[ 484.943726] env[62965]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=62965) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 484.944072] env[62965]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=62965) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 484.944197] env[62965]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=62965) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 484.944535] env[62965]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 485.040790] env[62965]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=62965) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 485.050701] env[62965]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=62965) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 485.663207] env[62965]: INFO nova.virt.driver [None req-efd4bd26-798d-40e9-80e7-c7a461ca4a88 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 485.734487] env[62965]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 485.734663] env[62965]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 485.734752] env[62965]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=62965) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 488.872830] env[62965]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-05fc5d57-0fc6-4b57-8a0d-da01247de952 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 488.889264] env[62965]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=62965) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 488.889457] env[62965]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-3a00d98a-d7b2-479a-b84e-b5a17376fd66 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 488.923342] env[62965]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 7aa1c. [ 488.923924] env[62965]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.189s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 488.924077] env[62965]: INFO nova.virt.vmwareapi.driver [None req-efd4bd26-798d-40e9-80e7-c7a461ca4a88 None None] VMware vCenter version: 7.0.3 [ 488.927496] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-074c5af9-7e46-472f-aa4f-4a9ccb67dfd8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 488.944573] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f66b3712-1f8a-4cef-b4dc-ec9930d73862 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 488.950467] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eda6950-2276-4afa-86bb-fa351d94a48d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 488.957590] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4dff700-8cb0-4263-9f35-9e09004ecb49 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 488.970461] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d07fd3f6-b2b0-4722-a36a-ee91c56daa6f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 488.976276] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5409eb6-8804-48c7-b263-f97fd8a2e789 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 489.006193] env[62965]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-eafb3fbf-c235-4e66-967c-c98e2a7a9c73 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 489.010956] env[62965]: DEBUG nova.virt.vmwareapi.driver [None req-efd4bd26-798d-40e9-80e7-c7a461ca4a88 None None] Extension org.openstack.compute already exists. {{(pid=62965) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:228}} [ 489.013600] env[62965]: INFO nova.compute.provider_config [None req-efd4bd26-798d-40e9-80e7-c7a461ca4a88 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 489.517260] env[62965]: DEBUG nova.context [None req-efd4bd26-798d-40e9-80e7-c7a461ca4a88 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),e5ad046b-b19a-4fd7-8f98-35ff1275dc90(cell1) {{(pid=62965) load_cells /opt/stack/nova/nova/context.py:464}} [ 489.519452] env[62965]: DEBUG oslo_concurrency.lockutils [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 489.519682] env[62965]: DEBUG oslo_concurrency.lockutils [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 489.520422] env[62965]: DEBUG oslo_concurrency.lockutils [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 489.520860] env[62965]: DEBUG oslo_concurrency.lockutils [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] Acquiring lock "e5ad046b-b19a-4fd7-8f98-35ff1275dc90" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 489.521209] env[62965]: DEBUG oslo_concurrency.lockutils [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] Lock "e5ad046b-b19a-4fd7-8f98-35ff1275dc90" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 489.522154] env[62965]: DEBUG oslo_concurrency.lockutils [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] Lock "e5ad046b-b19a-4fd7-8f98-35ff1275dc90" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 489.544506] env[62965]: INFO dbcounter [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] Registered counter for database nova_cell0 [ 489.552711] env[62965]: INFO dbcounter [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] Registered counter for database nova_cell1 [ 489.555996] env[62965]: DEBUG oslo_db.sqlalchemy.engines [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62965) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 489.556368] env[62965]: DEBUG oslo_db.sqlalchemy.engines [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62965) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 489.562141] env[62965]: ERROR nova.db.main.api [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 489.562141] env[62965]: result = function(*args, **kwargs) [ 489.562141] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 489.562141] env[62965]: return func(*args, **kwargs) [ 489.562141] env[62965]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 489.562141] env[62965]: result = fn(*args, **kwargs) [ 489.562141] env[62965]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 489.562141] env[62965]: return f(*args, **kwargs) [ 489.562141] env[62965]: File "/opt/stack/nova/nova/objects/service.py", line 556, in _db_service_get_minimum_version [ 489.562141] env[62965]: return db.service_get_minimum_version(context, binaries) [ 489.562141] env[62965]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 489.562141] env[62965]: _check_db_access() [ 489.562141] env[62965]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 489.562141] env[62965]: stacktrace = ''.join(traceback.format_stack()) [ 489.562141] env[62965]: [ 489.563116] env[62965]: ERROR nova.db.main.api [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 489.563116] env[62965]: result = function(*args, **kwargs) [ 489.563116] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 489.563116] env[62965]: return func(*args, **kwargs) [ 489.563116] env[62965]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 489.563116] env[62965]: result = fn(*args, **kwargs) [ 489.563116] env[62965]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 489.563116] env[62965]: return f(*args, **kwargs) [ 489.563116] env[62965]: File "/opt/stack/nova/nova/objects/service.py", line 556, in _db_service_get_minimum_version [ 489.563116] env[62965]: return db.service_get_minimum_version(context, binaries) [ 489.563116] env[62965]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 489.563116] env[62965]: _check_db_access() [ 489.563116] env[62965]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 489.563116] env[62965]: stacktrace = ''.join(traceback.format_stack()) [ 489.563116] env[62965]: [ 489.563529] env[62965]: WARNING nova.objects.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 489.563667] env[62965]: WARNING nova.objects.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] Failed to get minimum service version for cell e5ad046b-b19a-4fd7-8f98-35ff1275dc90 [ 489.564104] env[62965]: DEBUG oslo_concurrency.lockutils [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] Acquiring lock "singleton_lock" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 489.564273] env[62965]: DEBUG oslo_concurrency.lockutils [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] Acquired lock "singleton_lock" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 489.564521] env[62965]: DEBUG oslo_concurrency.lockutils [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] Releasing lock "singleton_lock" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 489.564835] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] Full set of CONF: {{(pid=62965) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 489.564982] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ******************************************************************************** {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 489.565268] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] Configuration options gathered from: {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 489.565268] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 489.565495] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 489.565602] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ================================================================================ {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 489.565865] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] allow_resize_to_same_host = True {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.565984] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] arq_binding_timeout = 300 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.566143] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] backdoor_port = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.566274] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] backdoor_socket = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.566447] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] block_device_allocate_retries = 60 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.566612] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] block_device_allocate_retries_interval = 3 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.566785] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cert = self.pem {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.566964] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.567178] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] compute_monitors = [] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.567356] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] config_dir = [] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.567529] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] config_drive_format = iso9660 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.567664] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.567830] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] config_source = [] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.568006] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] console_host = devstack {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.568186] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] control_exchange = nova {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.568348] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cpu_allocation_ratio = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.568511] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] daemon = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.568682] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] debug = True {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.568841] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] default_access_ip_network_name = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.569015] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] default_availability_zone = nova {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.569182] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] default_ephemeral_format = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.569345] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] default_green_pool_size = 1000 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.569581] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.569746] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] default_schedule_zone = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.569907] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] disk_allocation_ratio = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.570086] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] enable_new_services = True {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.570271] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] enabled_apis = ['osapi_compute'] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.570439] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] enabled_ssl_apis = [] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.570602] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] flat_injected = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.570761] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] force_config_drive = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.570925] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] force_raw_images = True {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.571111] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] graceful_shutdown_timeout = 5 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.571277] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] heal_instance_info_cache_interval = 60 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.571489] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] host = cpu-1 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.571669] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] initial_cpu_allocation_ratio = 4.0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.571836] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] initial_disk_allocation_ratio = 1.0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.572007] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] initial_ram_allocation_ratio = 1.0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.572230] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.572397] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] instance_build_timeout = 0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.572560] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] instance_delete_interval = 300 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.572730] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] instance_format = [instance: %(uuid)s] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.572898] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] instance_name_template = instance-%08x {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.573081] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] instance_usage_audit = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.573255] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] instance_usage_audit_period = month {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.573423] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.573589] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] instances_path = /opt/stack/data/nova/instances {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.573756] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] internal_service_availability_zone = internal {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.573916] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] key = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.574095] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] live_migration_retry_count = 30 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.574268] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] log_color = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.574433] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] log_config_append = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.574600] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.574760] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] log_dir = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.574920] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] log_file = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.575065] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] log_options = True {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.575233] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] log_rotate_interval = 1 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.575431] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] log_rotate_interval_type = days {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.575621] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] log_rotation_type = none {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.575754] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.575908] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.576103] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.576279] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.576409] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.576575] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] long_rpc_timeout = 1800 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.576741] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] max_concurrent_builds = 10 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.576908] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] max_concurrent_live_migrations = 1 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.577113] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] max_concurrent_snapshots = 5 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.577289] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] max_local_block_devices = 3 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.577453] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] max_logfile_count = 30 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.577615] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] max_logfile_size_mb = 200 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.577777] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] maximum_instance_delete_attempts = 5 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.577947] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] metadata_listen = 0.0.0.0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.578136] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] metadata_listen_port = 8775 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.578303] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] metadata_workers = 2 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.578466] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] migrate_max_retries = -1 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.578685] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] mkisofs_cmd = genisoimage {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.578923] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] my_block_storage_ip = 10.180.1.21 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.579076] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] my_ip = 10.180.1.21 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.579291] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] my_shared_fs_storage_ip = 10.180.1.21 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.579461] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] network_allocate_retries = 0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.579649] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.579819] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] osapi_compute_listen = 0.0.0.0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.579987] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] osapi_compute_listen_port = 8774 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.580178] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] osapi_compute_unique_server_name_scope = {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.580351] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] osapi_compute_workers = 2 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.580516] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] password_length = 12 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.580680] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] periodic_enable = True {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.580842] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] periodic_fuzzy_delay = 60 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.581015] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] pointer_model = usbtablet {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.581191] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] preallocate_images = none {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.581355] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] publish_errors = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.581486] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] pybasedir = /opt/stack/nova {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.581648] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ram_allocation_ratio = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.581809] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] rate_limit_burst = 0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.581978] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] rate_limit_except_level = CRITICAL {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.582158] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] rate_limit_interval = 0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.582346] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] reboot_timeout = 0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.582514] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] reclaim_instance_interval = 0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.582677] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] record = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.582849] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] reimage_timeout_per_gb = 60 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.583043] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] report_interval = 120 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.583277] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] rescue_timeout = 0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.583465] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] reserved_host_cpus = 0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.583634] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] reserved_host_disk_mb = 0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.583799] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] reserved_host_memory_mb = 512 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.583988] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] reserved_huge_pages = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.584179] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] resize_confirm_window = 0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.584344] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] resize_fs_using_block_device = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.584506] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] resume_guests_state_on_host_boot = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.584680] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.584840] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] rpc_response_timeout = 60 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.585056] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] run_external_periodic_tasks = True {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.585239] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] running_deleted_instance_action = reap {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.585409] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] running_deleted_instance_poll_interval = 1800 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.585573] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] running_deleted_instance_timeout = 0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.585737] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] scheduler_instance_sync_interval = 120 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.585940] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] service_down_time = 720 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.586146] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] servicegroup_driver = db {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.586314] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] shell_completion = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.586482] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] shelved_offload_time = 0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.586677] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] shelved_poll_interval = 3600 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.586873] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] shutdown_timeout = 0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.587075] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] source_is_ipv6 = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.587251] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ssl_only = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.587508] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.587684] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] sync_power_state_interval = 600 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.587852] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] sync_power_state_pool_size = 1000 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.588033] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] syslog_log_facility = LOG_USER {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.588200] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] tempdir = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.588363] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] timeout_nbd = 10 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.588532] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] transport_url = **** {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.588699] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] update_resources_interval = 0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.588863] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] use_cow_images = True {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.589040] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] use_eventlog = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.589207] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] use_journal = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.589370] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] use_json = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.589535] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] use_rootwrap_daemon = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.589698] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] use_stderr = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.589862] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] use_syslog = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.590033] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vcpu_pin_set = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.590207] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vif_plugging_is_fatal = True {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.590378] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vif_plugging_timeout = 300 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.590549] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] virt_mkfs = [] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.590716] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] volume_usage_poll_interval = 0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.590882] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] watch_log_file = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.591068] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] web = /usr/share/spice-html5 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 489.591263] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.591434] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.591601] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] os_brick.wait_mpath_device_interval = 1 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.591778] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_concurrency.disable_process_locking = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.592089] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.592298] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.592477] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.592655] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_metrics.metrics_process_name = {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.592838] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.593022] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.593208] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api.auth_strategy = keystone {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.593383] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api.compute_link_prefix = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.593563] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.593745] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api.dhcp_domain = novalocal {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.593940] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api.enable_instance_password = True {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.594139] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api.glance_link_prefix = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.594315] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.594493] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api.instance_list_cells_batch_strategy = distributed {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.594663] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api.instance_list_per_project_cells = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.594829] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api.list_records_by_skipping_down_cells = True {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.594999] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api.local_metadata_per_cell = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.595190] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api.max_limit = 1000 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.595364] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api.metadata_cache_expiration = 15 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.595544] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api.neutron_default_tenant_id = default {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.595723] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api.response_validation = warn {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.595920] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api.use_neutron_default_nets = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.596126] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.596298] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api.vendordata_dynamic_failure_fatal = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.596473] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.596650] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api.vendordata_dynamic_ssl_certfile = {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.596824] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api.vendordata_dynamic_targets = [] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.596993] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api.vendordata_jsonfile_path = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.597197] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api.vendordata_providers = ['StaticJSON'] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.597395] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.backend = dogpile.cache.memcached {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.597567] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.backend_argument = **** {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.597743] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.config_prefix = cache.oslo {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.597916] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.dead_timeout = 60.0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.598103] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.debug_cache_backend = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.598297] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.enable_retry_client = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.598474] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.enable_socket_keepalive = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.598652] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.enabled = True {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.598822] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.enforce_fips_mode = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.598996] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.expiration_time = 600 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.599178] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.hashclient_retry_attempts = 2 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.599349] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.hashclient_retry_delay = 1.0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.599517] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.memcache_dead_retry = 300 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.599679] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.memcache_password = **** {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.599847] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.600028] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.600203] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.memcache_pool_maxsize = 10 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.600369] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.memcache_pool_unused_timeout = 60 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.600535] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.memcache_sasl_enabled = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.600716] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.memcache_servers = ['localhost:11211'] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.600888] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.memcache_socket_timeout = 1.0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.601067] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.memcache_username = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.601245] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.proxies = [] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.601417] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.redis_db = 0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.601577] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.redis_password = **** {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.601748] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.redis_sentinel_service_name = mymaster {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.601927] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.602117] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.redis_server = localhost:6379 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.602288] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.redis_socket_timeout = 1.0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.602451] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.redis_username = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.602617] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.retry_attempts = 2 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.602785] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.retry_delay = 0.0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.602952] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.socket_keepalive_count = 1 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.603136] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.socket_keepalive_idle = 1 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.603302] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.socket_keepalive_interval = 1 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.603464] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.tls_allowed_ciphers = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.603628] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.tls_cafile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.603786] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.tls_certfile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.603973] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.tls_enabled = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.604161] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cache.tls_keyfile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.604339] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cinder.auth_section = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.604519] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cinder.auth_type = password {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.604686] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cinder.cafile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.604906] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cinder.catalog_info = volumev3::publicURL {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.605655] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cinder.certfile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.605655] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cinder.collect_timing = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.605655] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cinder.cross_az_attach = True {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.605655] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cinder.debug = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.605798] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cinder.endpoint_template = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.605881] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cinder.http_retries = 3 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.606095] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cinder.insecure = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.606268] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cinder.keyfile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.606448] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cinder.os_region_name = RegionOne {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.606618] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cinder.split_loggers = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.606783] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cinder.timeout = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.606961] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.607143] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] compute.cpu_dedicated_set = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.607308] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] compute.cpu_shared_set = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.607476] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] compute.image_type_exclude_list = [] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.607642] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] compute.live_migration_wait_for_vif_plug = True {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.607809] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] compute.max_concurrent_disk_ops = 0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.607978] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] compute.max_disk_devices_to_attach = -1 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.608163] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.608356] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.608525] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] compute.resource_provider_association_refresh = 300 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.608691] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.608859] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] compute.shutdown_retry_interval = 10 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.609055] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.609245] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] conductor.workers = 2 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.609429] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] console.allowed_origins = [] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.609594] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] console.ssl_ciphers = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.609773] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] console.ssl_minimum_version = default {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.609946] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] consoleauth.enforce_session_timeout = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.610136] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] consoleauth.token_ttl = 600 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.610310] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cyborg.cafile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.610473] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cyborg.certfile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.610640] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cyborg.collect_timing = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.610806] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cyborg.connect_retries = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.610971] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cyborg.connect_retry_delay = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.611150] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cyborg.endpoint_override = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.611318] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cyborg.insecure = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.611479] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cyborg.keyfile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.611644] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cyborg.max_version = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.611804] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cyborg.min_version = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.611965] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cyborg.region_name = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.612145] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cyborg.retriable_status_codes = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.612308] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cyborg.service_name = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.612482] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cyborg.service_type = accelerator {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.612645] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cyborg.split_loggers = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.612808] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cyborg.status_code_retries = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.612971] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cyborg.status_code_retry_delay = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.613151] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cyborg.timeout = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.613337] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.613505] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] cyborg.version = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.613688] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] database.backend = sqlalchemy {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.613873] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] database.connection = **** {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.614080] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] database.connection_debug = 0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.614264] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] database.connection_parameters = {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.614450] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] database.connection_recycle_time = 3600 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.614634] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] database.connection_trace = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.614815] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] database.db_inc_retry_interval = True {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.615008] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] database.db_max_retries = 20 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.615194] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] database.db_max_retry_interval = 10 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.615380] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] database.db_retry_interval = 1 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.615563] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] database.max_overflow = 50 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.615747] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] database.max_pool_size = 5 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.615953] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] database.max_retries = 10 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.616162] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] database.mysql_sql_mode = TRADITIONAL {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.616337] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] database.mysql_wsrep_sync_wait = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.616508] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] database.pool_timeout = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.616679] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] database.retry_interval = 10 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.616846] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] database.slave_connection = **** {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.617018] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] database.sqlite_synchronous = True {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.617195] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] database.use_db_reconnect = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.617382] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api_database.backend = sqlalchemy {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.617557] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api_database.connection = **** {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.617728] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api_database.connection_debug = 0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.617904] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api_database.connection_parameters = {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.618088] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api_database.connection_recycle_time = 3600 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.618283] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api_database.connection_trace = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.618460] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api_database.db_inc_retry_interval = True {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.618630] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api_database.db_max_retries = 20 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.618801] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api_database.db_max_retry_interval = 10 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.618971] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api_database.db_retry_interval = 1 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.619155] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api_database.max_overflow = 50 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.619327] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api_database.max_pool_size = 5 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.619495] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api_database.max_retries = 10 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.619669] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.619835] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api_database.mysql_wsrep_sync_wait = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.620006] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api_database.pool_timeout = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.620185] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api_database.retry_interval = 10 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.620349] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api_database.slave_connection = **** {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.620517] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] api_database.sqlite_synchronous = True {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.620697] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] devices.enabled_mdev_types = [] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.620881] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.621070] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ephemeral_storage_encryption.default_format = luks {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.621241] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ephemeral_storage_encryption.enabled = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.621410] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ephemeral_storage_encryption.key_size = 512 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.621585] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] glance.api_servers = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.621756] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] glance.cafile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.621922] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] glance.certfile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.622103] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] glance.collect_timing = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.622271] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] glance.connect_retries = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.622435] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] glance.connect_retry_delay = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.622601] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] glance.debug = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.622770] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] glance.default_trusted_certificate_ids = [] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.622937] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] glance.enable_certificate_validation = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.623119] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] glance.enable_rbd_download = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.623286] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] glance.endpoint_override = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.623457] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] glance.insecure = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.623623] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] glance.keyfile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.623786] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] glance.max_version = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.623974] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] glance.min_version = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.624169] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] glance.num_retries = 3 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.624343] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] glance.rbd_ceph_conf = {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.624512] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] glance.rbd_connect_timeout = 5 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.624683] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] glance.rbd_pool = {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.624856] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] glance.rbd_user = {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.625114] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] glance.region_name = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.625196] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] glance.retriable_status_codes = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.625353] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] glance.service_name = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.625524] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] glance.service_type = image {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.625696] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] glance.split_loggers = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.625879] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] glance.status_code_retries = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.626061] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] glance.status_code_retry_delay = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.626233] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] glance.timeout = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.626420] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.626588] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] glance.verify_glance_signatures = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.626752] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] glance.version = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.626924] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] guestfs.debug = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.627115] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] manila.auth_section = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.627283] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] manila.auth_type = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.627446] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] manila.cafile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.627606] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] manila.certfile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.627772] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] manila.collect_timing = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.627937] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] manila.connect_retries = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.628116] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] manila.connect_retry_delay = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.628305] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] manila.endpoint_override = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.628479] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] manila.insecure = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.628640] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] manila.keyfile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.628803] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] manila.max_version = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.628965] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] manila.min_version = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.629145] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] manila.region_name = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.629311] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] manila.retriable_status_codes = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.629471] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] manila.service_name = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.629642] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] manila.service_type = shared-file-system {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.629810] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] manila.share_apply_policy_timeout = 10 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.629980] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] manila.split_loggers = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.630158] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] manila.status_code_retries = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.630325] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] manila.status_code_retry_delay = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.630491] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] manila.timeout = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.630677] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] manila.valid_interfaces = ['internal', 'public'] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.630844] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] manila.version = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.631026] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] mks.enabled = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.631398] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.631594] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] image_cache.manager_interval = 2400 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.631771] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] image_cache.precache_concurrency = 1 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.631948] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] image_cache.remove_unused_base_images = True {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.632139] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.632314] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.632496] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] image_cache.subdirectory_name = _base {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.632674] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ironic.api_max_retries = 60 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.632843] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ironic.api_retry_interval = 2 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.633018] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ironic.auth_section = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.633194] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ironic.auth_type = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.633358] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ironic.cafile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.633521] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ironic.certfile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.633689] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ironic.collect_timing = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.633867] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ironic.conductor_group = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.634062] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ironic.connect_retries = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.634234] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ironic.connect_retry_delay = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.634398] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ironic.endpoint_override = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.634564] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ironic.insecure = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.634725] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ironic.keyfile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.634886] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ironic.max_version = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.635057] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ironic.min_version = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.635233] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ironic.peer_list = [] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.635395] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ironic.region_name = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.635558] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ironic.retriable_status_codes = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.635726] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ironic.serial_console_state_timeout = 10 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.635912] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ironic.service_name = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.636111] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ironic.service_type = baremetal {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.636280] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ironic.shard = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.636446] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ironic.split_loggers = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.636610] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ironic.status_code_retries = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.636775] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ironic.status_code_retry_delay = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.636937] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ironic.timeout = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.637138] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.637307] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ironic.version = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.637493] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.637672] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] key_manager.fixed_key = **** {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.637865] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.638038] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] barbican.barbican_api_version = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.638223] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] barbican.barbican_endpoint = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.638409] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] barbican.barbican_endpoint_type = public {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.638574] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] barbican.barbican_region_name = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.638738] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] barbican.cafile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.639202] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] barbican.certfile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.639202] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] barbican.collect_timing = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.639272] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] barbican.insecure = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.639385] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] barbican.keyfile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.639553] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] barbican.number_of_retries = 60 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.639718] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] barbican.retry_delay = 1 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.639883] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] barbican.send_service_user_token = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.640060] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] barbican.split_loggers = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.640226] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] barbican.timeout = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.640388] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] barbican.verify_ssl = True {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.640547] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] barbican.verify_ssl_path = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.640718] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] barbican_service_user.auth_section = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.640886] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] barbican_service_user.auth_type = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.641072] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] barbican_service_user.cafile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.641250] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] barbican_service_user.certfile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.641418] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] barbican_service_user.collect_timing = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.641580] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] barbican_service_user.insecure = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.641742] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] barbican_service_user.keyfile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.641906] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] barbican_service_user.split_loggers = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.642081] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] barbican_service_user.timeout = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.642254] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vault.approle_role_id = **** {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.642414] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vault.approle_secret_id = **** {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.642586] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vault.kv_mountpoint = secret {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.642746] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vault.kv_path = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.642910] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vault.kv_version = 2 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.643082] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vault.namespace = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.643246] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vault.root_token_id = **** {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.643408] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vault.ssl_ca_crt_file = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.643574] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vault.timeout = 60.0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.643738] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vault.use_ssl = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.643929] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.644129] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] keystone.auth_section = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.644301] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] keystone.auth_type = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.644464] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] keystone.cafile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.644623] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] keystone.certfile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.644787] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] keystone.collect_timing = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.644950] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] keystone.connect_retries = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.645139] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] keystone.connect_retry_delay = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.645304] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] keystone.endpoint_override = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.645457] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] keystone.insecure = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.645616] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] keystone.keyfile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.645774] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] keystone.max_version = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.645978] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] keystone.min_version = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.646201] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] keystone.region_name = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.646375] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] keystone.retriable_status_codes = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.646539] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] keystone.service_name = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.646714] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] keystone.service_type = identity {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.646884] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] keystone.split_loggers = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.647061] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] keystone.status_code_retries = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.647230] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] keystone.status_code_retry_delay = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.647394] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] keystone.timeout = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.647578] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.647743] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] keystone.version = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.647950] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.connection_uri = {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.648135] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.cpu_mode = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.648339] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.cpu_model_extra_flags = [] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.648520] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.cpu_models = [] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.648696] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.cpu_power_governor_high = performance {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.648869] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.cpu_power_governor_low = powersave {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.649050] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.cpu_power_management = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.649230] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.649400] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.device_detach_attempts = 8 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.649567] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.device_detach_timeout = 20 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.649737] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.disk_cachemodes = [] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.649901] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.disk_prefix = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.650096] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.enabled_perf_events = [] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.650274] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.file_backed_memory = 0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.650449] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.gid_maps = [] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.650613] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.hw_disk_discard = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.650776] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.hw_machine_type = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.650948] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.images_rbd_ceph_conf = {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.651133] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.651300] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.651472] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.images_rbd_glance_store_name = {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.651642] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.images_rbd_pool = rbd {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.651813] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.images_type = default {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.651976] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.images_volume_group = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.652157] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.inject_key = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.652326] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.inject_partition = -2 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.652491] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.inject_password = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.652655] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.iscsi_iface = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.652819] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.iser_use_multipath = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.652987] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.live_migration_bandwidth = 0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.653170] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.live_migration_completion_timeout = 800 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.653340] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.live_migration_downtime = 500 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.653505] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.live_migration_downtime_delay = 75 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.653670] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.live_migration_downtime_steps = 10 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.653833] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.live_migration_inbound_addr = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.654035] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.live_migration_permit_auto_converge = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.654212] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.live_migration_permit_post_copy = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.654374] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.live_migration_scheme = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.654546] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.live_migration_timeout_action = abort {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.654712] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.live_migration_tunnelled = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.654875] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.live_migration_uri = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.655053] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.live_migration_with_native_tls = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.655219] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.max_queues = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.655400] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.mem_stats_period_seconds = 10 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.655624] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.655789] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.nfs_mount_options = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.656129] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.656316] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.num_aoe_discover_tries = 3 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.656489] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.num_iser_scan_tries = 5 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.656656] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.num_memory_encrypted_guests = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.656826] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.num_nvme_discover_tries = 5 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.657013] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.num_pcie_ports = 0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.657204] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.num_volume_scan_tries = 5 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.657378] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.pmem_namespaces = [] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.657542] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.quobyte_client_cfg = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.657839] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.658028] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.rbd_connect_timeout = 5 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.658229] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.658440] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.658614] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.rbd_secret_uuid = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.658780] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.rbd_user = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.658950] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.realtime_scheduler_priority = 1 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.659142] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.remote_filesystem_transport = ssh {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.659311] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.rescue_image_id = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.659473] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.rescue_kernel_id = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.659637] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.rescue_ramdisk_id = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.659810] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.rng_dev_path = /dev/urandom {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.659973] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.rx_queue_size = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.660162] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.smbfs_mount_options = {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.660444] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.660620] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.snapshot_compression = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.660785] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.snapshot_image_format = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.661011] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.661194] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.sparse_logical_volumes = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.661361] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.swtpm_enabled = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.661535] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.swtpm_group = tss {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.661707] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.swtpm_user = tss {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.661880] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.sysinfo_serial = unique {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.662064] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.tb_cache_size = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.662229] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.tx_queue_size = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.662397] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.uid_maps = [] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.662562] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.use_virtio_for_bridges = True {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.662733] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.virt_type = kvm {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.662905] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.volume_clear = zero {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.663087] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.volume_clear_size = 0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.663260] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.volume_use_multipath = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.663423] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.vzstorage_cache_path = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.663599] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.663772] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.vzstorage_mount_group = qemu {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.663991] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.vzstorage_mount_opts = [] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.664191] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.664477] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.664658] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.vzstorage_mount_user = stack {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.664828] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.665065] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] neutron.auth_section = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.665288] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] neutron.auth_type = password {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.665491] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] neutron.cafile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.665624] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] neutron.certfile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.665792] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] neutron.collect_timing = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.665986] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] neutron.connect_retries = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.666174] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] neutron.connect_retry_delay = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.666349] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] neutron.default_floating_pool = public {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.666516] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] neutron.endpoint_override = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.666688] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] neutron.extension_sync_interval = 600 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.666863] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] neutron.http_retries = 3 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.667039] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] neutron.insecure = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.667210] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] neutron.keyfile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.667375] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] neutron.max_version = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.667548] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] neutron.metadata_proxy_shared_secret = **** {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.667711] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] neutron.min_version = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.667882] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] neutron.ovs_bridge = br-int {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.668065] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] neutron.physnets = [] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.668243] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] neutron.region_name = RegionOne {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.668410] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] neutron.retriable_status_codes = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.668583] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] neutron.service_metadata_proxy = True {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.668747] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] neutron.service_name = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.668919] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] neutron.service_type = network {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.669101] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] neutron.split_loggers = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.669270] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] neutron.status_code_retries = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.669433] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] neutron.status_code_retry_delay = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.669598] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] neutron.timeout = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.669783] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.669953] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] neutron.version = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.670145] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] notifications.bdms_in_notifications = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.670330] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] notifications.default_level = INFO {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.670509] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] notifications.notification_format = unversioned {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.670677] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] notifications.notify_on_state_change = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.670857] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.671052] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] pci.alias = [] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.671232] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] pci.device_spec = [] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.671402] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] pci.report_in_placement = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.671577] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.auth_section = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.671753] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.auth_type = password {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.671924] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.auth_url = http://10.180.1.21/identity {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.672110] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.cafile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.672278] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.certfile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.672445] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.collect_timing = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.672610] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.connect_retries = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.672772] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.connect_retry_delay = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.672936] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.default_domain_id = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.673118] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.default_domain_name = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.673277] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.domain_id = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.673438] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.domain_name = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.673598] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.endpoint_override = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.673765] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.insecure = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.673943] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.keyfile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.674129] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.max_version = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.674293] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.min_version = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.674465] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.password = **** {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.674628] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.project_domain_id = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.674796] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.project_domain_name = Default {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.674967] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.project_id = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.675157] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.project_name = service {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.675330] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.region_name = RegionOne {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.675503] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.retriable_status_codes = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.675659] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.service_name = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.675825] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.service_type = placement {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.676032] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.split_loggers = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.676211] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.status_code_retries = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.676377] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.status_code_retry_delay = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.676539] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.system_scope = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.676699] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.timeout = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.676863] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.trust_id = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.677047] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.user_domain_id = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.677226] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.user_domain_name = Default {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.677390] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.user_id = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.677566] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.username = nova {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.677750] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.677914] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] placement.version = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.678110] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] quota.cores = 20 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.678281] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] quota.count_usage_from_placement = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.678455] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.678635] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] quota.injected_file_content_bytes = 10240 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.678808] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] quota.injected_file_path_length = 255 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.678984] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] quota.injected_files = 5 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.679172] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] quota.instances = 10 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.679345] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] quota.key_pairs = 100 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.679514] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] quota.metadata_items = 128 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.679708] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] quota.ram = 51200 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.679926] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] quota.recheck_quota = True {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.680124] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] quota.server_group_members = 10 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.680303] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] quota.server_groups = 10 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.680483] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.680653] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.680821] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] scheduler.image_metadata_prefilter = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.680987] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.681174] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] scheduler.max_attempts = 3 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.681344] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] scheduler.max_placement_results = 1000 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.681512] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.681679] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] scheduler.query_placement_for_image_type_support = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.681845] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.682029] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] scheduler.workers = 2 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.682215] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.682390] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.682574] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.682746] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.682916] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.683096] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.683268] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.683461] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.683635] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] filter_scheduler.host_subset_size = 1 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.683834] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.684042] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] filter_scheduler.image_properties_default_architecture = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.684225] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.684398] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] filter_scheduler.isolated_hosts = [] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.684567] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] filter_scheduler.isolated_images = [] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.684733] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] filter_scheduler.max_instances_per_host = 50 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.684899] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.685079] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.685248] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] filter_scheduler.pci_in_placement = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.685414] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.685578] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.685745] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.685953] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.686147] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.686318] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.686484] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] filter_scheduler.track_instance_changes = True {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.686665] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.686837] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] metrics.required = True {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.687010] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] metrics.weight_multiplier = 1.0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.687186] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] metrics.weight_of_unavailable = -10000.0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.687354] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] metrics.weight_setting = [] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.687675] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.687854] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] serial_console.enabled = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.688046] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] serial_console.port_range = 10000:20000 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.688228] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.688401] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.688572] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] serial_console.serialproxy_port = 6083 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.688745] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] service_user.auth_section = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.688927] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] service_user.auth_type = password {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.689117] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] service_user.cafile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.689286] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] service_user.certfile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.689452] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] service_user.collect_timing = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.689617] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] service_user.insecure = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.689797] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] service_user.keyfile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.690054] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] service_user.send_service_user_token = True {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.690245] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] service_user.split_loggers = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.690415] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] service_user.timeout = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.690592] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] spice.agent_enabled = True {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.690762] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] spice.enabled = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.691094] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.691300] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] spice.html5proxy_host = 0.0.0.0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.691476] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] spice.html5proxy_port = 6082 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.691643] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] spice.image_compression = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.691809] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] spice.jpeg_compression = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.691973] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] spice.playback_compression = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.692163] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] spice.require_secure = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.692336] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] spice.server_listen = 127.0.0.1 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.692508] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.692673] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] spice.streaming_mode = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.692833] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] spice.zlib_compression = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.693016] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] upgrade_levels.baseapi = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.693200] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] upgrade_levels.compute = auto {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.693368] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] upgrade_levels.conductor = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.693533] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] upgrade_levels.scheduler = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.693703] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vendordata_dynamic_auth.auth_section = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.693910] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vendordata_dynamic_auth.auth_type = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.694108] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vendordata_dynamic_auth.cafile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.694279] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vendordata_dynamic_auth.certfile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.694450] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vendordata_dynamic_auth.collect_timing = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.694611] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vendordata_dynamic_auth.insecure = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.694773] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vendordata_dynamic_auth.keyfile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.694938] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vendordata_dynamic_auth.split_loggers = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.695115] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vendordata_dynamic_auth.timeout = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.695332] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vmware.api_retry_count = 10 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.695505] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vmware.ca_file = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.695682] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vmware.cache_prefix = devstack-image-cache {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.695896] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vmware.cluster_name = testcl1 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.696104] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vmware.connection_pool_size = 10 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.696280] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vmware.console_delay_seconds = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.696455] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vmware.datastore_regex = ^datastore.* {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.696667] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.696849] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vmware.host_password = **** {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.697071] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vmware.host_port = 443 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.697261] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vmware.host_username = administrator@vsphere.local {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.697459] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vmware.insecure = True {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.697656] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vmware.integration_bridge = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.697830] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vmware.maximum_objects = 100 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.697995] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vmware.pbm_default_policy = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.698182] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vmware.pbm_enabled = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.698347] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vmware.pbm_wsdl_location = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.698520] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.698684] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vmware.serial_port_proxy_uri = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.698846] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vmware.serial_port_service_uri = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.699027] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vmware.task_poll_interval = 0.5 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.699238] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vmware.use_linked_clone = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.699422] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vmware.vnc_keymap = en-us {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.699598] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vmware.vnc_port = 5900 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.699788] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vmware.vnc_port_total = 10000 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.699983] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vnc.auth_schemes = ['none'] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.700180] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vnc.enabled = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.700487] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.700677] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.700852] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vnc.novncproxy_port = 6080 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.701045] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vnc.server_listen = 127.0.0.1 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.701232] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.701399] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vnc.vencrypt_ca_certs = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.701561] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vnc.vencrypt_client_cert = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.701723] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vnc.vencrypt_client_key = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.701909] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.702091] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] workarounds.disable_deep_image_inspection = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.702259] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] workarounds.disable_fallback_pcpu_query = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.702450] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] workarounds.disable_group_policy_check_upcall = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.702660] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.702836] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] workarounds.disable_rootwrap = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.703009] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] workarounds.enable_numa_live_migration = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.703249] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.703438] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.703611] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] workarounds.handle_virt_lifecycle_events = True {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.703877] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] workarounds.libvirt_disable_apic = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.704161] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] workarounds.never_download_image_if_on_rbd = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.704427] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.704709] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.705020] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.705322] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.705604] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.705913] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.706144] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.706339] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.706517] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.706710] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.706891] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] wsgi.client_socket_timeout = 900 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.707080] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] wsgi.default_pool_size = 1000 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.707258] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] wsgi.keep_alive = True {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.707432] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] wsgi.max_header_line = 16384 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.707600] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] wsgi.secure_proxy_ssl_header = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.707769] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] wsgi.ssl_ca_file = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.707935] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] wsgi.ssl_cert_file = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.708115] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] wsgi.ssl_key_file = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.708288] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] wsgi.tcp_keepidle = 600 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.708471] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.708643] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] zvm.ca_file = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.708810] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] zvm.cloud_connector_url = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.709503] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.709700] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] zvm.reachable_timeout = 300 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.709896] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_policy.enforce_new_defaults = True {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.710299] env[62965]: WARNING oslo_config.cfg [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 489.710495] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_policy.enforce_scope = True {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.710680] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_policy.policy_default_rule = default {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.710874] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.711070] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_policy.policy_file = policy.yaml {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.711256] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.711426] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.711590] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.711754] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.711921] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.712109] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.712293] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.712474] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] profiler.connection_string = messaging:// {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.712645] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] profiler.enabled = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.712819] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] profiler.es_doc_type = notification {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.712988] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] profiler.es_scroll_size = 10000 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.713178] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] profiler.es_scroll_time = 2m {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.713350] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] profiler.filter_error_trace = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.713522] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] profiler.hmac_keys = **** {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.713693] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] profiler.sentinel_service_name = mymaster {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.713875] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] profiler.socket_timeout = 0.1 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.714072] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] profiler.trace_requests = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.714245] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] profiler.trace_sqlalchemy = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.714435] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] profiler_jaeger.process_tags = {} {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.714602] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] profiler_jaeger.service_name_prefix = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.714769] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] profiler_otlp.service_name_prefix = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.714940] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] remote_debug.host = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.715121] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] remote_debug.port = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.715311] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.715481] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.715652] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.715820] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.716018] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.716195] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.716363] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.716529] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.716696] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.716875] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_rabbit.hostname = devstack {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.717049] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.717229] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.717401] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.717576] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.717748] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.717922] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.718103] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.718285] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.718455] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.718622] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.718791] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.718961] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.719145] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.719318] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.719485] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.719651] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.719816] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.719980] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.720170] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.720340] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_rabbit.ssl = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.720516] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.720689] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.720856] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.721042] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.721219] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_rabbit.ssl_version = {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.721386] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.721577] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.721751] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_notifications.retry = -1 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.721938] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.722131] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_messaging_notifications.transport_url = **** {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.722312] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_limit.auth_section = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.722482] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_limit.auth_type = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.722644] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_limit.cafile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.722808] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_limit.certfile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.722973] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_limit.collect_timing = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.723151] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_limit.connect_retries = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.723315] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_limit.connect_retry_delay = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.723480] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_limit.endpoint_id = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.723641] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_limit.endpoint_override = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.723807] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_limit.insecure = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.723995] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_limit.keyfile = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.724180] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_limit.max_version = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.724346] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_limit.min_version = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.724508] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_limit.region_name = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.724675] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_limit.retriable_status_codes = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.724840] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_limit.service_name = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.725008] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_limit.service_type = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.725185] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_limit.split_loggers = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.725352] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_limit.status_code_retries = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.725516] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_limit.status_code_retry_delay = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.725680] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_limit.timeout = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.725852] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_limit.valid_interfaces = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.726067] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_limit.version = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.726228] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_reports.file_event_handler = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.726395] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_reports.file_event_handler_interval = 1 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.726563] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] oslo_reports.log_dir = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.726738] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.726919] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vif_plug_linux_bridge_privileged.group = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.727131] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.727311] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.727482] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.727659] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vif_plug_linux_bridge_privileged.user = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.727834] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.727997] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vif_plug_ovs_privileged.group = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.728175] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vif_plug_ovs_privileged.helper_command = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.728345] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.728511] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.728672] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] vif_plug_ovs_privileged.user = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.728846] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] os_vif_linux_bridge.flat_interface = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.729095] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.729300] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.729483] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.729659] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.729837] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.730018] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.730193] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] os_vif_linux_bridge.vlan_interface = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.730378] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.730555] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] os_vif_ovs.isolate_vif = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.730728] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.730899] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.731091] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.731264] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] os_vif_ovs.ovsdb_interface = native {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.731430] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] os_vif_ovs.per_port_bridge = False {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.731607] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] privsep_osbrick.capabilities = [21] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.731770] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] privsep_osbrick.group = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.731932] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] privsep_osbrick.helper_command = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.732119] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.732289] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] privsep_osbrick.thread_pool_size = 8 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.732453] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] privsep_osbrick.user = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.732632] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.732796] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] nova_sys_admin.group = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.732958] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] nova_sys_admin.helper_command = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.733143] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.733314] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] nova_sys_admin.thread_pool_size = 8 {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.733475] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] nova_sys_admin.user = None {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 489.733606] env[62965]: DEBUG oslo_service.service [None req-055b6000-e060-4bb0-9c09-c58dd4e4239e None None] ******************************************************************************** {{(pid=62965) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 489.734132] env[62965]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 490.237522] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-2f681fb2-9321-4efc-a090-2e3f4f3f408e None None] Getting list of instances from cluster (obj){ [ 490.237522] env[62965]: value = "domain-c8" [ 490.237522] env[62965]: _type = "ClusterComputeResource" [ 490.237522] env[62965]: } {{(pid=62965) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 490.238709] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5faa4e6-3e8a-4b16-8deb-174b387a5a31 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 490.247583] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-2f681fb2-9321-4efc-a090-2e3f4f3f408e None None] Got total of 0 instances {{(pid=62965) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 490.248113] env[62965]: WARNING nova.virt.vmwareapi.driver [None req-2f681fb2-9321-4efc-a090-2e3f4f3f408e None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 490.248595] env[62965]: INFO nova.virt.node [None req-2f681fb2-9321-4efc-a090-2e3f4f3f408e None None] Generated node identity 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 [ 490.248839] env[62965]: INFO nova.virt.node [None req-2f681fb2-9321-4efc-a090-2e3f4f3f408e None None] Wrote node identity 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 to /opt/stack/data/n-cpu-1/compute_id [ 490.751321] env[62965]: WARNING nova.compute.manager [None req-2f681fb2-9321-4efc-a090-2e3f4f3f408e None None] Compute nodes ['20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 491.756267] env[62965]: INFO nova.compute.manager [None req-2f681fb2-9321-4efc-a090-2e3f4f3f408e None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 492.762127] env[62965]: WARNING nova.compute.manager [None req-2f681fb2-9321-4efc-a090-2e3f4f3f408e None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 492.762574] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2f681fb2-9321-4efc-a090-2e3f4f3f408e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 492.762669] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2f681fb2-9321-4efc-a090-2e3f4f3f408e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 492.762748] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2f681fb2-9321-4efc-a090-2e3f4f3f408e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 492.762907] env[62965]: DEBUG nova.compute.resource_tracker [None req-2f681fb2-9321-4efc-a090-2e3f4f3f408e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62965) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 492.763842] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a5488f2-94ea-485c-adc3-a1f960472995 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.772176] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f7ec038-c432-4ad7-ae55-f18a0f0232a7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.787500] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33f05a98-557b-4366-9ac9-654ff23fc08b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.794168] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34055369-7d11-4566-ad93-01e18414dbad {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.823538] env[62965]: DEBUG nova.compute.resource_tracker [None req-2f681fb2-9321-4efc-a090-2e3f4f3f408e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181375MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62965) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 492.823714] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2f681fb2-9321-4efc-a090-2e3f4f3f408e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 492.823883] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2f681fb2-9321-4efc-a090-2e3f4f3f408e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 493.326716] env[62965]: WARNING nova.compute.resource_tracker [None req-2f681fb2-9321-4efc-a090-2e3f4f3f408e None None] No compute node record for cpu-1:20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 could not be found. [ 493.830021] env[62965]: INFO nova.compute.resource_tracker [None req-2f681fb2-9321-4efc-a090-2e3f4f3f408e None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 [ 495.337360] env[62965]: DEBUG nova.compute.resource_tracker [None req-2f681fb2-9321-4efc-a090-2e3f4f3f408e None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62965) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 495.337803] env[62965]: DEBUG nova.compute.resource_tracker [None req-2f681fb2-9321-4efc-a090-2e3f4f3f408e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62965) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 495.487482] env[62965]: INFO nova.scheduler.client.report [None req-2f681fb2-9321-4efc-a090-2e3f4f3f408e None None] [req-3373d8c1-c400-44ad-aa52-737135fbd271] Created resource provider record via placement API for resource provider with UUID 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 495.502944] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-becdef95-c69d-4208-be86-16a59cdbeaa7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 495.510861] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43fa7d5e-d509-4a2a-bed9-2dee6431aaf8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 495.541245] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12151aa1-82d5-44b8-b992-353b302d80b7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 495.548735] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1972244b-4de9-4ce1-bb3a-81041a128f34 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 495.561755] env[62965]: DEBUG nova.compute.provider_tree [None req-2f681fb2-9321-4efc-a090-2e3f4f3f408e None None] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 496.096514] env[62965]: DEBUG nova.scheduler.client.report [None req-2f681fb2-9321-4efc-a090-2e3f4f3f408e None None] Updated inventory for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 496.096745] env[62965]: DEBUG nova.compute.provider_tree [None req-2f681fb2-9321-4efc-a090-2e3f4f3f408e None None] Updating resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 generation from 0 to 1 during operation: update_inventory {{(pid=62965) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 496.096885] env[62965]: DEBUG nova.compute.provider_tree [None req-2f681fb2-9321-4efc-a090-2e3f4f3f408e None None] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 496.146847] env[62965]: DEBUG nova.compute.provider_tree [None req-2f681fb2-9321-4efc-a090-2e3f4f3f408e None None] Updating resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 generation from 1 to 2 during operation: update_traits {{(pid=62965) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 496.651999] env[62965]: DEBUG nova.compute.resource_tracker [None req-2f681fb2-9321-4efc-a090-2e3f4f3f408e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62965) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 496.652442] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2f681fb2-9321-4efc-a090-2e3f4f3f408e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.828s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 496.652442] env[62965]: DEBUG nova.service [None req-2f681fb2-9321-4efc-a090-2e3f4f3f408e None None] Creating RPC server for service compute {{(pid=62965) start /opt/stack/nova/nova/service.py:186}} [ 496.665071] env[62965]: DEBUG nova.service [None req-2f681fb2-9321-4efc-a090-2e3f4f3f408e None None] Join ServiceGroup membership for this service compute {{(pid=62965) start /opt/stack/nova/nova/service.py:203}} [ 496.665248] env[62965]: DEBUG nova.servicegroup.drivers.db [None req-2f681fb2-9321-4efc-a090-2e3f4f3f408e None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=62965) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 516.668588] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._sync_power_states {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 517.176028] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Getting list of instances from cluster (obj){ [ 517.176028] env[62965]: value = "domain-c8" [ 517.176028] env[62965]: _type = "ClusterComputeResource" [ 517.176028] env[62965]: } {{(pid=62965) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 517.176028] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-183da930-4c35-4189-877f-50137b608301 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.183037] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Got total of 0 instances {{(pid=62965) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 517.183455] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 517.183896] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Getting list of instances from cluster (obj){ [ 517.183896] env[62965]: value = "domain-c8" [ 517.183896] env[62965]: _type = "ClusterComputeResource" [ 517.183896] env[62965]: } {{(pid=62965) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 517.184996] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c7d9d05-e020-47f0-853e-8f5d0963bbef {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.192226] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Got total of 0 instances {{(pid=62965) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 536.268249] env[62965]: DEBUG oslo_concurrency.lockutils [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Acquiring lock "532ae225-da05-4cbb-9bb4-c72a8451acff" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.269233] env[62965]: DEBUG oslo_concurrency.lockutils [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Lock "532ae225-da05-4cbb-9bb4-c72a8451acff" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 536.771548] env[62965]: DEBUG nova.compute.manager [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 537.316103] env[62965]: DEBUG oslo_concurrency.lockutils [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.316103] env[62965]: DEBUG oslo_concurrency.lockutils [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 537.316782] env[62965]: INFO nova.compute.claims [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 538.495144] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72c27b6e-b476-4c2e-b2e2-52def2f7de42 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.508380] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce9db724-0dbd-4f39-9192-1022236f7148 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.543700] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abe7afa2-6378-4394-8e75-32ffe3c0f9c6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.552567] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a09c0e5-cef9-461d-ad51-133686d03d6c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.570664] env[62965]: DEBUG nova.compute.provider_tree [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 538.803905] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] Acquiring lock "585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.804451] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] Lock "585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.074213] env[62965]: DEBUG nova.scheduler.client.report [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 539.309733] env[62965]: DEBUG nova.compute.manager [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 539.578864] env[62965]: DEBUG oslo_concurrency.lockutils [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.264s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 539.581315] env[62965]: DEBUG nova.compute.manager [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 539.802055] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Acquiring lock "ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.802332] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Lock "ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.836114] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.836384] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.838492] env[62965]: INFO nova.compute.claims [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 540.055663] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] Acquiring lock "1e128a17-cc1a-4aa0-8020-9c01775d9f75" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.055663] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] Lock "1e128a17-cc1a-4aa0-8020-9c01775d9f75" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.088205] env[62965]: DEBUG nova.compute.utils [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 540.093792] env[62965]: DEBUG nova.compute.manager [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Not allocating networking since 'none' was specified. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 540.306671] env[62965]: DEBUG nova.compute.manager [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 540.559562] env[62965]: DEBUG nova.compute.manager [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 540.595598] env[62965]: DEBUG nova.compute.manager [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 540.849414] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.964326] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f132218b-ca65-4c7e-8ff4-3b9877122e10 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.972852] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-986f3dde-2410-4ee2-8468-4e3009526f62 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.011145] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86f002cc-07d3-401c-97a4-af66ca4cb1c9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.020759] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9e7dc28-a224-435f-a925-a02d4978ad73 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.043312] env[62965]: DEBUG nova.compute.provider_tree [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 541.088729] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.279121] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Acquiring lock "c46f89de-5156-49b4-b8f7-f60d78f1134b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.279499] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Lock "c46f89de-5156-49b4-b8f7-f60d78f1134b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.547319] env[62965]: DEBUG nova.scheduler.client.report [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 541.610961] env[62965]: DEBUG nova.compute.manager [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 541.786402] env[62965]: DEBUG nova.compute.manager [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 541.854067] env[62965]: DEBUG nova.virt.hardware [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 541.855136] env[62965]: DEBUG nova.virt.hardware [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 541.855136] env[62965]: DEBUG nova.virt.hardware [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 541.855586] env[62965]: DEBUG nova.virt.hardware [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 541.856262] env[62965]: DEBUG nova.virt.hardware [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 541.856502] env[62965]: DEBUG nova.virt.hardware [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 541.856944] env[62965]: DEBUG nova.virt.hardware [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 541.857258] env[62965]: DEBUG nova.virt.hardware [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 541.858285] env[62965]: DEBUG nova.virt.hardware [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 541.858417] env[62965]: DEBUG nova.virt.hardware [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 541.858811] env[62965]: DEBUG nova.virt.hardware [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 541.860684] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b28da5b-6822-4812-889f-f9cee4483111 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.875059] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfcd98ea-3563-4098-8d89-9fddaaf121a4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.896440] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-444e312c-6baa-49bb-8823-8fa4a32d2f15 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.927105] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Instance VIF info [] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 541.936810] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 541.937213] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4ba2e0cd-860e-41aa-82a5-f1513e469ca1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.951871] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Created folder: OpenStack in parent group-v4. [ 541.952257] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Creating folder: Project (f4f2a08f90194913a4b968190ee77503). Parent ref: group-v294931. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 541.956078] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-494754f0-4e40-41e5-8714-d6a9cad88cbe {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.963114] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Created folder: Project (f4f2a08f90194913a4b968190ee77503) in parent group-v294931. [ 541.963228] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Creating folder: Instances. Parent ref: group-v294932. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 541.963495] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-13569460-bc72-4d0c-8a6a-68e89758c13c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.975216] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Created folder: Instances in parent group-v294932. [ 541.975475] env[62965]: DEBUG oslo.service.loopingcall [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 541.975663] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 541.975914] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9c593d8a-84b1-4196-b5bb-286b809b198f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.998071] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 541.998071] env[62965]: value = "task-1389717" [ 541.998071] env[62965]: _type = "Task" [ 541.998071] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 542.012160] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389717, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 542.053856] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.217s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 542.053942] env[62965]: DEBUG nova.compute.manager [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 542.060705] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.210s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.060705] env[62965]: INFO nova.compute.claims [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 542.325461] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.488109] env[62965]: DEBUG oslo_concurrency.lockutils [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] Acquiring lock "709bac33-9605-4254-bf38-49fc7892b15e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.488109] env[62965]: DEBUG oslo_concurrency.lockutils [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] Lock "709bac33-9605-4254-bf38-49fc7892b15e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.003s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.522566] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389717, 'name': CreateVM_Task, 'duration_secs': 0.321793} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 542.523706] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 542.526622] env[62965]: DEBUG oslo_vmware.service [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98d9778a-6cf7-4747-b33d-5d76cf44e47b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.534263] env[62965]: DEBUG oslo_concurrency.lockutils [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 542.534432] env[62965]: DEBUG oslo_concurrency.lockutils [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 542.535376] env[62965]: DEBUG oslo_concurrency.lockutils [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 542.535731] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1881c0ee-d857-4e51-b17c-94691374aa0c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.541369] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Waiting for the task: (returnval){ [ 542.541369] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52fc8bd0-1b95-d1f4-5c32-ab4d53d1ed0b" [ 542.541369] env[62965]: _type = "Task" [ 542.541369] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 542.556299] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52fc8bd0-1b95-d1f4-5c32-ab4d53d1ed0b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 542.569190] env[62965]: DEBUG nova.compute.utils [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 542.571800] env[62965]: DEBUG nova.compute.manager [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 542.571800] env[62965]: DEBUG nova.network.neutron [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 542.783956] env[62965]: DEBUG nova.policy [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6670f79125484fdcaf2a8dfef01158bb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '29d38da1c41f4e1eb3b123c1314a6002', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 542.826327] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] Acquiring lock "7780b346-28a4-4d80-9f83-a452c23741ec" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.826431] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] Lock "7780b346-28a4-4d80-9f83-a452c23741ec" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.991233] env[62965]: DEBUG nova.compute.manager [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 543.061154] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52fc8bd0-1b95-d1f4-5c32-ab4d53d1ed0b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.076842] env[62965]: DEBUG nova.compute.manager [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 543.224537] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82d2a922-515b-4a7a-ab6d-91bda1837ee9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.233205] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7723b668-95c9-4081-952c-c0a456b570cb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.271924] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7a3b7b6-19b7-4a67-a288-7ab67b7e6e8f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.280380] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c051ffb1-1860-4add-b62d-5dc519112b28 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.298279] env[62965]: DEBUG nova.compute.provider_tree [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 543.330428] env[62965]: DEBUG nova.compute.manager [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 543.527378] env[62965]: DEBUG oslo_concurrency.lockutils [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.558932] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52fc8bd0-1b95-d1f4-5c32-ab4d53d1ed0b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.561609] env[62965]: DEBUG nova.network.neutron [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Successfully created port: c2c98b22-9f4c-4cdf-9b90-5856a23c4cce {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 543.802789] env[62965]: DEBUG nova.scheduler.client.report [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 543.853993] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.068525] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52fc8bd0-1b95-d1f4-5c32-ab4d53d1ed0b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 544.090949] env[62965]: DEBUG nova.compute.manager [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 544.141387] env[62965]: DEBUG nova.virt.hardware [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 544.141905] env[62965]: DEBUG nova.virt.hardware [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 544.141905] env[62965]: DEBUG nova.virt.hardware [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 544.142088] env[62965]: DEBUG nova.virt.hardware [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 544.142232] env[62965]: DEBUG nova.virt.hardware [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 544.142375] env[62965]: DEBUG nova.virt.hardware [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 544.142573] env[62965]: DEBUG nova.virt.hardware [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 544.142723] env[62965]: DEBUG nova.virt.hardware [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 544.142882] env[62965]: DEBUG nova.virt.hardware [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 544.143137] env[62965]: DEBUG nova.virt.hardware [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 544.143211] env[62965]: DEBUG nova.virt.hardware [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 544.144463] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c252d5f2-48f9-4548-b371-5a41151d11de {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.157094] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27ab6cf0-d5bb-4b78-98af-b10838b23e2e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.196911] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Acquiring lock "fa55e796-81d0-46b7-b025-9119bae2f305" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.197315] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Lock "fa55e796-81d0-46b7-b025-9119bae2f305" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.308851] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.250s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 544.309464] env[62965]: DEBUG nova.compute.manager [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 544.312342] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.223s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.313736] env[62965]: INFO nova.compute.claims [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 544.559850] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52fc8bd0-1b95-d1f4-5c32-ab4d53d1ed0b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 544.706130] env[62965]: DEBUG nova.compute.manager [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 544.823861] env[62965]: DEBUG nova.compute.utils [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 544.828474] env[62965]: DEBUG nova.compute.manager [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 544.828474] env[62965]: DEBUG nova.network.neutron [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 544.968349] env[62965]: DEBUG nova.policy [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '96279b4d04cb46949ccc343b4a6198da', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '63609764674145849796434c49d2299c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 545.064121] env[62965]: DEBUG oslo_concurrency.lockutils [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 545.065253] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 545.065253] env[62965]: DEBUG oslo_concurrency.lockutils [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 545.065725] env[62965]: DEBUG oslo_concurrency.lockutils [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 545.066315] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 545.067063] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-eb14b281-2657-4ca7-a1ed-d92a414c7a7b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.086680] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 545.086875] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 545.087696] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eda1a10f-e08f-44a1-b842-513943ff9385 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.099639] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f553466-a041-4375-8da4-5c821fae7d09 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.106220] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Waiting for the task: (returnval){ [ 545.106220] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]527cb412-cd70-747d-529d-78fcd4114afa" [ 545.106220] env[62965]: _type = "Task" [ 545.106220] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 545.113125] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 545.113550] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 545.113652] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Starting heal instance info cache {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10257}} [ 545.113953] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Rebuilding the list of instances to heal {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10261}} [ 545.127295] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Preparing fetch location {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 545.127722] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Creating directory with path [datastore1] vmware_temp/68ceacce-77f6-4eae-bc2c-36c55ccfd2dd/523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 545.128107] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d06450ab-a6a7-4c7a-a0e2-88461245409b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.152251] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Created directory with path [datastore1] vmware_temp/68ceacce-77f6-4eae-bc2c-36c55ccfd2dd/523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 545.153103] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Fetch image to [datastore1] vmware_temp/68ceacce-77f6-4eae-bc2c-36c55ccfd2dd/523fe138-8131-44f9-96b3-87f923f60cb8/tmp-sparse.vmdk {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 545.153103] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Downloading image file data 523fe138-8131-44f9-96b3-87f923f60cb8 to [datastore1] vmware_temp/68ceacce-77f6-4eae-bc2c-36c55ccfd2dd/523fe138-8131-44f9-96b3-87f923f60cb8/tmp-sparse.vmdk on the data store datastore1 {{(pid=62965) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 545.153563] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dae02ff3-5498-4bd8-a68a-d3ca15064c10 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.165850] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ac1e0bb-435f-4b28-8697-2073b0c231ea {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.180797] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3abb153-36e3-4c50-9905-1d6f1e31f59d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.238148] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50d89c15-537e-4006-b53d-8e8d2097adea {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.248737] env[62965]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-b1c9da55-28a0-435d-82e5-51b54932d930 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.264754] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.328325] env[62965]: DEBUG nova.compute.manager [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 545.345235] env[62965]: DEBUG nova.virt.vmwareapi.images [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Downloading image file data 523fe138-8131-44f9-96b3-87f923f60cb8 to the data store datastore1 {{(pid=62965) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 545.443332] env[62965]: DEBUG oslo_vmware.rw_handles [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/68ceacce-77f6-4eae-bc2c-36c55ccfd2dd/523fe138-8131-44f9-96b3-87f923f60cb8/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62965) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 545.622422] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Skipping network cache update for instance because it is Building. {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10270}} [ 545.622592] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Skipping network cache update for instance because it is Building. {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10270}} [ 545.622718] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Skipping network cache update for instance because it is Building. {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10270}} [ 545.622844] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Skipping network cache update for instance because it is Building. {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10270}} [ 545.622964] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Didn't find any instances for network info cache update. {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10343}} [ 545.627844] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 545.628151] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 545.629022] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 545.629022] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 545.629022] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 545.629022] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 545.629302] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62965) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10876}} [ 545.629302] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 545.634941] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2564c8f6-d5bb-4e19-be86-0d51a3053cf2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.650753] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Acquiring lock "b811920a-14a2-4c11-8b13-78d920f08a0e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.650753] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Lock "b811920a-14a2-4c11-8b13-78d920f08a0e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.652975] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cb1af53-4d0a-47c9-9528-600cd74daab5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.711140] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40c30552-d9b6-4b4b-8d17-0c2b7383541c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.715689] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Acquiring lock "4a75ddc4-add2-45dc-b086-bc163990db4f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.715923] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Lock "4a75ddc4-add2-45dc-b086-bc163990db4f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.731849] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e23c7e75-1bf5-4301-85b2-32f0c745f2fb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.737762] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Acquiring lock "2870d8cb-8c19-454e-8a6e-1675eba8ddd4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.737892] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Lock "2870d8cb-8c19-454e-8a6e-1675eba8ddd4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.751615] env[62965]: DEBUG nova.compute.provider_tree [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 546.110096] env[62965]: DEBUG oslo_vmware.rw_handles [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Completed reading data from the image iterator. {{(pid=62965) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 546.110096] env[62965]: DEBUG oslo_vmware.rw_handles [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Closing write handle for https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/68ceacce-77f6-4eae-bc2c-36c55ccfd2dd/523fe138-8131-44f9-96b3-87f923f60cb8/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62965) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 546.135463] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.152957] env[62965]: DEBUG nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 546.221519] env[62965]: DEBUG nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 546.260082] env[62965]: DEBUG nova.scheduler.client.report [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 546.264220] env[62965]: DEBUG nova.virt.vmwareapi.images [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Downloaded image file data 523fe138-8131-44f9-96b3-87f923f60cb8 to vmware_temp/68ceacce-77f6-4eae-bc2c-36c55ccfd2dd/523fe138-8131-44f9-96b3-87f923f60cb8/tmp-sparse.vmdk on the data store datastore1 {{(pid=62965) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 546.265974] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Caching image {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 546.266233] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Copying Virtual Disk [datastore1] vmware_temp/68ceacce-77f6-4eae-bc2c-36c55ccfd2dd/523fe138-8131-44f9-96b3-87f923f60cb8/tmp-sparse.vmdk to [datastore1] vmware_temp/68ceacce-77f6-4eae-bc2c-36c55ccfd2dd/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 546.267041] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f30d85a8-1b88-437a-a57c-132e385be4a2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.281546] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Waiting for the task: (returnval){ [ 546.281546] env[62965]: value = "task-1389718" [ 546.281546] env[62965]: _type = "Task" [ 546.281546] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 546.290901] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Task: {'id': task-1389718, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.336225] env[62965]: DEBUG nova.network.neutron [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Successfully created port: 0a2a87c1-e174-4aa2-b17c-54f557ffd83b {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 546.347243] env[62965]: DEBUG nova.compute.manager [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 546.386751] env[62965]: DEBUG nova.virt.hardware [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 546.386751] env[62965]: DEBUG nova.virt.hardware [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 546.387303] env[62965]: DEBUG nova.virt.hardware [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 546.387303] env[62965]: DEBUG nova.virt.hardware [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 546.387303] env[62965]: DEBUG nova.virt.hardware [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 546.387396] env[62965]: DEBUG nova.virt.hardware [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 546.389367] env[62965]: DEBUG nova.virt.hardware [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 546.389367] env[62965]: DEBUG nova.virt.hardware [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 546.389367] env[62965]: DEBUG nova.virt.hardware [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 546.391124] env[62965]: DEBUG nova.virt.hardware [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 546.391124] env[62965]: DEBUG nova.virt.hardware [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 546.391204] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1cd1880-758e-4642-88b5-7b749fd6e9a2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.405858] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b077ac52-d30d-48aa-888c-500f3f79658a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.682935] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.754742] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.773099] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.461s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 546.773546] env[62965]: DEBUG nova.compute.manager [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 546.776234] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.451s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.782779] env[62965]: INFO nova.compute.claims [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 546.793131] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Task: {'id': task-1389718, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 547.157263] env[62965]: ERROR nova.compute.manager [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c2c98b22-9f4c-4cdf-9b90-5856a23c4cce, please check neutron logs for more information. [ 547.157263] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 547.157263] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 547.157263] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 547.157263] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 547.157263] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 547.157263] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 547.157263] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 547.157263] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.157263] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 547.157263] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.157263] env[62965]: ERROR nova.compute.manager raise self.value [ 547.157263] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 547.157263] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 547.157263] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.157263] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 547.158260] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.158260] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 547.158260] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c2c98b22-9f4c-4cdf-9b90-5856a23c4cce, please check neutron logs for more information. [ 547.158260] env[62965]: ERROR nova.compute.manager [ 547.158814] env[62965]: Traceback (most recent call last): [ 547.158889] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 547.158889] env[62965]: listener.cb(fileno) [ 547.158889] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 547.158889] env[62965]: result = function(*args, **kwargs) [ 547.158889] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 547.158889] env[62965]: return func(*args, **kwargs) [ 547.158889] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 547.158889] env[62965]: raise e [ 547.158889] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 547.158889] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 547.158889] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 547.158889] env[62965]: created_port_ids = self._update_ports_for_instance( [ 547.158889] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 547.158889] env[62965]: with excutils.save_and_reraise_exception(): [ 547.158889] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.158889] env[62965]: self.force_reraise() [ 547.158889] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.158889] env[62965]: raise self.value [ 547.158889] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 547.158889] env[62965]: updated_port = self._update_port( [ 547.158889] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.158889] env[62965]: _ensure_no_port_binding_failure(port) [ 547.158889] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.158889] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 547.158889] env[62965]: nova.exception.PortBindingFailed: Binding failed for port c2c98b22-9f4c-4cdf-9b90-5856a23c4cce, please check neutron logs for more information. [ 547.159801] env[62965]: Removing descriptor: 16 [ 547.162016] env[62965]: ERROR nova.compute.manager [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c2c98b22-9f4c-4cdf-9b90-5856a23c4cce, please check neutron logs for more information. [ 547.162016] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Traceback (most recent call last): [ 547.162016] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 547.162016] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] yield resources [ 547.162016] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 547.162016] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] self.driver.spawn(context, instance, image_meta, [ 547.162016] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 547.162016] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 547.162016] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 547.162016] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] vm_ref = self.build_virtual_machine(instance, [ 547.162016] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 547.162558] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] vif_infos = vmwarevif.get_vif_info(self._session, [ 547.162558] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 547.162558] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] for vif in network_info: [ 547.162558] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 547.162558] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] return self._sync_wrapper(fn, *args, **kwargs) [ 547.162558] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 547.162558] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] self.wait() [ 547.162558] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 547.162558] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] self[:] = self._gt.wait() [ 547.162558] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 547.162558] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] return self._exit_event.wait() [ 547.162558] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 547.162558] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] result = hub.switch() [ 547.163035] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 547.163035] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] return self.greenlet.switch() [ 547.163035] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 547.163035] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] result = function(*args, **kwargs) [ 547.163035] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 547.163035] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] return func(*args, **kwargs) [ 547.163035] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 547.163035] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] raise e [ 547.163035] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 547.163035] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] nwinfo = self.network_api.allocate_for_instance( [ 547.163035] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 547.163035] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] created_port_ids = self._update_ports_for_instance( [ 547.163035] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 547.163713] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] with excutils.save_and_reraise_exception(): [ 547.163713] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.163713] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] self.force_reraise() [ 547.163713] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.163713] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] raise self.value [ 547.163713] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 547.163713] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] updated_port = self._update_port( [ 547.163713] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.163713] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] _ensure_no_port_binding_failure(port) [ 547.163713] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.163713] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] raise exception.PortBindingFailed(port_id=port['id']) [ 547.163713] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] nova.exception.PortBindingFailed: Binding failed for port c2c98b22-9f4c-4cdf-9b90-5856a23c4cce, please check neutron logs for more information. [ 547.163713] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] [ 547.164114] env[62965]: INFO nova.compute.manager [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Terminating instance [ 547.282109] env[62965]: DEBUG nova.compute.utils [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 547.286019] env[62965]: DEBUG nova.compute.manager [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 547.286249] env[62965]: DEBUG nova.network.neutron [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 547.301940] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Task: {'id': task-1389718, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.782007} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 547.303598] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Copied Virtual Disk [datastore1] vmware_temp/68ceacce-77f6-4eae-bc2c-36c55ccfd2dd/523fe138-8131-44f9-96b3-87f923f60cb8/tmp-sparse.vmdk to [datastore1] vmware_temp/68ceacce-77f6-4eae-bc2c-36c55ccfd2dd/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 547.304241] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Deleting the datastore file [datastore1] vmware_temp/68ceacce-77f6-4eae-bc2c-36c55ccfd2dd/523fe138-8131-44f9-96b3-87f923f60cb8/tmp-sparse.vmdk {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 547.304241] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2623a67b-c035-4622-bdbd-de75a6d74b9d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.314110] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Waiting for the task: (returnval){ [ 547.314110] env[62965]: value = "task-1389719" [ 547.314110] env[62965]: _type = "Task" [ 547.314110] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 547.327099] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Task: {'id': task-1389719, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 547.389070] env[62965]: DEBUG nova.policy [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '69c9a5c1cba046478cd2b4b8fe00f995', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4e527311aca14e419b63ddf277220fda', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 547.670686] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] Acquiring lock "refresh_cache-585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 547.670859] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] Acquired lock "refresh_cache-585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 547.671043] env[62965]: DEBUG nova.network.neutron [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 547.787597] env[62965]: DEBUG nova.compute.manager [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 547.816943] env[62965]: DEBUG nova.network.neutron [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Successfully created port: 52a9bae7-4ed9-425c-908c-82c0d761c688 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 547.834847] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Task: {'id': task-1389719, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.025598} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 547.835436] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 547.835663] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Moving file from [datastore1] vmware_temp/68ceacce-77f6-4eae-bc2c-36c55ccfd2dd/523fe138-8131-44f9-96b3-87f923f60cb8 to [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8. {{(pid=62965) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 547.835929] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-14227446-252f-45ec-b0af-d8b2fab388ec {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.853099] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Waiting for the task: (returnval){ [ 547.853099] env[62965]: value = "task-1389720" [ 547.853099] env[62965]: _type = "Task" [ 547.853099] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 547.862116] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Task: {'id': task-1389720, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 548.217700] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88fc4999-e260-4fa2-884d-11fee30688aa {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.223947] env[62965]: DEBUG nova.network.neutron [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 548.230179] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4e18998-42fb-44db-bcb8-b063a2722616 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.273991] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd4e5276-159e-411b-a89f-c8f7e8e761c7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.282646] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3de5be69-f8fd-4ba9-a924-d0fbd7d9d77d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.301386] env[62965]: DEBUG nova.compute.provider_tree [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 548.366861] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Task: {'id': task-1389720, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.036132} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 548.367611] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] File moved {{(pid=62965) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 548.367611] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Cleaning up location [datastore1] vmware_temp/68ceacce-77f6-4eae-bc2c-36c55ccfd2dd {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 548.367611] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Deleting the datastore file [datastore1] vmware_temp/68ceacce-77f6-4eae-bc2c-36c55ccfd2dd {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 548.367824] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-01eedd18-4806-4959-85fa-c818bb95adaa {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.374335] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Waiting for the task: (returnval){ [ 548.374335] env[62965]: value = "task-1389721" [ 548.374335] env[62965]: _type = "Task" [ 548.374335] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 548.382720] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Task: {'id': task-1389721, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 548.392566] env[62965]: DEBUG nova.network.neutron [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.803034] env[62965]: DEBUG nova.compute.manager [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 548.807676] env[62965]: DEBUG nova.scheduler.client.report [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 548.845491] env[62965]: DEBUG nova.virt.hardware [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 548.845874] env[62965]: DEBUG nova.virt.hardware [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 548.846145] env[62965]: DEBUG nova.virt.hardware [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 548.846563] env[62965]: DEBUG nova.virt.hardware [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 548.846811] env[62965]: DEBUG nova.virt.hardware [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 548.847073] env[62965]: DEBUG nova.virt.hardware [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 548.847951] env[62965]: DEBUG nova.virt.hardware [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 548.847951] env[62965]: DEBUG nova.virt.hardware [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 548.847951] env[62965]: DEBUG nova.virt.hardware [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 548.848186] env[62965]: DEBUG nova.virt.hardware [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 548.848425] env[62965]: DEBUG nova.virt.hardware [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 548.849814] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edc72cbc-9172-47eb-9555-ca77f73686b9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.858850] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41132811-fd8b-4d54-ab6c-105daa35c76a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.887026] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Task: {'id': task-1389721, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.025809} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 548.887223] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 548.888023] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f0f1f78-ceb8-49b1-b59c-9a7c8eab49b2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.895283] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] Releasing lock "refresh_cache-585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 548.895283] env[62965]: DEBUG nova.compute.manager [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 548.895283] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 548.895283] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Waiting for the task: (returnval){ [ 548.895283] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]521e029e-84cb-d313-9389-af667e543728" [ 548.895283] env[62965]: _type = "Task" [ 548.895283] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 548.895834] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-013d4e50-c0f9-465e-b6ea-ca49b1e37225 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.912198] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d9c322f-717c-4fee-9c0c-d883fc2a5d75 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.934354] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]521e029e-84cb-d313-9389-af667e543728, 'name': SearchDatastore_Task, 'duration_secs': 0.009727} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 548.935081] env[62965]: DEBUG oslo_concurrency.lockutils [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 548.935344] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] 532ae225-da05-4cbb-9bb4-c72a8451acff/532ae225-da05-4cbb-9bb4-c72a8451acff.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 548.935596] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a9543edc-ecf8-4e4d-a1d1-ceb6831297ee {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.942380] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Waiting for the task: (returnval){ [ 548.942380] env[62965]: value = "task-1389722" [ 548.942380] env[62965]: _type = "Task" [ 548.942380] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 548.947736] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5 could not be found. [ 548.947846] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 548.948807] env[62965]: INFO nova.compute.manager [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Took 0.05 seconds to destroy the instance on the hypervisor. [ 548.948807] env[62965]: DEBUG oslo.service.loopingcall [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 548.952601] env[62965]: DEBUG nova.compute.manager [-] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 548.952601] env[62965]: DEBUG nova.network.neutron [-] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 548.959981] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Task: {'id': task-1389722, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 549.011534] env[62965]: DEBUG nova.network.neutron [-] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 549.313752] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.537s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 549.314385] env[62965]: DEBUG nova.compute.manager [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 549.318264] env[62965]: DEBUG oslo_concurrency.lockutils [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.790s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.318863] env[62965]: INFO nova.compute.claims [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 549.460580] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Task: {'id': task-1389722, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 549.514051] env[62965]: DEBUG nova.network.neutron [-] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.531434] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Acquiring lock "647a3ba2-8b84-45f9-9d28-130eb0521d1b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.531635] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Lock "647a3ba2-8b84-45f9-9d28-130eb0521d1b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.827280] env[62965]: DEBUG nova.compute.utils [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 549.833262] env[62965]: DEBUG nova.compute.manager [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 549.833708] env[62965]: DEBUG nova.network.neutron [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 549.959161] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Task: {'id': task-1389722, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.5252} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 549.959591] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] 532ae225-da05-4cbb-9bb4-c72a8451acff/532ae225-da05-4cbb-9bb4-c72a8451acff.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 549.959792] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 549.962327] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-86af74b2-8f28-4420-8745-b1ef4a2a2a8b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.966920] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Waiting for the task: (returnval){ [ 549.966920] env[62965]: value = "task-1389723" [ 549.966920] env[62965]: _type = "Task" [ 549.966920] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 549.975765] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Task: {'id': task-1389723, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 550.009801] env[62965]: DEBUG nova.policy [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a6788bd87356434cae74fe73701dd01e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'be2499f8fb014b9384c22dae0b09ed6b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 550.021213] env[62965]: INFO nova.compute.manager [-] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Took 1.07 seconds to deallocate network for instance. [ 550.025953] env[62965]: DEBUG nova.compute.claims [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 550.025953] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.138840] env[62965]: ERROR nova.compute.manager [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0a2a87c1-e174-4aa2-b17c-54f557ffd83b, please check neutron logs for more information. [ 550.138840] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 550.138840] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 550.138840] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 550.138840] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 550.138840] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 550.138840] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 550.138840] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 550.138840] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 550.138840] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 550.138840] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 550.138840] env[62965]: ERROR nova.compute.manager raise self.value [ 550.138840] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 550.138840] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 550.138840] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 550.138840] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 550.139407] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 550.139407] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 550.139407] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0a2a87c1-e174-4aa2-b17c-54f557ffd83b, please check neutron logs for more information. [ 550.139407] env[62965]: ERROR nova.compute.manager [ 550.139407] env[62965]: Traceback (most recent call last): [ 550.139407] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 550.139407] env[62965]: listener.cb(fileno) [ 550.139407] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 550.139407] env[62965]: result = function(*args, **kwargs) [ 550.139407] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 550.139407] env[62965]: return func(*args, **kwargs) [ 550.139407] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 550.139407] env[62965]: raise e [ 550.139407] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 550.139407] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 550.139407] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 550.139407] env[62965]: created_port_ids = self._update_ports_for_instance( [ 550.139407] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 550.139407] env[62965]: with excutils.save_and_reraise_exception(): [ 550.139407] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 550.139407] env[62965]: self.force_reraise() [ 550.139407] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 550.139407] env[62965]: raise self.value [ 550.139407] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 550.139407] env[62965]: updated_port = self._update_port( [ 550.139407] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 550.139407] env[62965]: _ensure_no_port_binding_failure(port) [ 550.139407] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 550.139407] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 550.140176] env[62965]: nova.exception.PortBindingFailed: Binding failed for port 0a2a87c1-e174-4aa2-b17c-54f557ffd83b, please check neutron logs for more information. [ 550.140176] env[62965]: Removing descriptor: 17 [ 550.140176] env[62965]: ERROR nova.compute.manager [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0a2a87c1-e174-4aa2-b17c-54f557ffd83b, please check neutron logs for more information. [ 550.140176] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Traceback (most recent call last): [ 550.140176] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 550.140176] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] yield resources [ 550.140176] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 550.140176] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] self.driver.spawn(context, instance, image_meta, [ 550.140176] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 550.140176] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 550.140176] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 550.140176] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] vm_ref = self.build_virtual_machine(instance, [ 550.140492] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 550.140492] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] vif_infos = vmwarevif.get_vif_info(self._session, [ 550.140492] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 550.140492] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] for vif in network_info: [ 550.140492] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 550.140492] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] return self._sync_wrapper(fn, *args, **kwargs) [ 550.140492] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 550.140492] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] self.wait() [ 550.140492] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 550.140492] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] self[:] = self._gt.wait() [ 550.140492] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 550.140492] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] return self._exit_event.wait() [ 550.140492] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 550.140830] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] result = hub.switch() [ 550.140830] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 550.140830] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] return self.greenlet.switch() [ 550.140830] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 550.140830] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] result = function(*args, **kwargs) [ 550.140830] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 550.140830] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] return func(*args, **kwargs) [ 550.140830] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 550.140830] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] raise e [ 550.140830] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 550.140830] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] nwinfo = self.network_api.allocate_for_instance( [ 550.140830] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 550.140830] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] created_port_ids = self._update_ports_for_instance( [ 550.141177] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 550.141177] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] with excutils.save_and_reraise_exception(): [ 550.141177] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 550.141177] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] self.force_reraise() [ 550.141177] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 550.141177] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] raise self.value [ 550.141177] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 550.141177] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] updated_port = self._update_port( [ 550.141177] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 550.141177] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] _ensure_no_port_binding_failure(port) [ 550.141177] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 550.141177] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] raise exception.PortBindingFailed(port_id=port['id']) [ 550.141751] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] nova.exception.PortBindingFailed: Binding failed for port 0a2a87c1-e174-4aa2-b17c-54f557ffd83b, please check neutron logs for more information. [ 550.141751] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] [ 550.141751] env[62965]: INFO nova.compute.manager [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Terminating instance [ 550.205183] env[62965]: DEBUG nova.compute.manager [req-6c57f27a-5181-4374-99f4-75aa53d35f2c req-0f826206-4c4d-4c3b-981a-cec9bff14732 service nova] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Received event network-changed-c2c98b22-9f4c-4cdf-9b90-5856a23c4cce {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 550.205439] env[62965]: DEBUG nova.compute.manager [req-6c57f27a-5181-4374-99f4-75aa53d35f2c req-0f826206-4c4d-4c3b-981a-cec9bff14732 service nova] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Refreshing instance network info cache due to event network-changed-c2c98b22-9f4c-4cdf-9b90-5856a23c4cce. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 550.205582] env[62965]: DEBUG oslo_concurrency.lockutils [req-6c57f27a-5181-4374-99f4-75aa53d35f2c req-0f826206-4c4d-4c3b-981a-cec9bff14732 service nova] Acquiring lock "refresh_cache-585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 550.205688] env[62965]: DEBUG oslo_concurrency.lockutils [req-6c57f27a-5181-4374-99f4-75aa53d35f2c req-0f826206-4c4d-4c3b-981a-cec9bff14732 service nova] Acquired lock "refresh_cache-585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 550.205845] env[62965]: DEBUG nova.network.neutron [req-6c57f27a-5181-4374-99f4-75aa53d35f2c req-0f826206-4c4d-4c3b-981a-cec9bff14732 service nova] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Refreshing network info cache for port c2c98b22-9f4c-4cdf-9b90-5856a23c4cce {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 550.334181] env[62965]: DEBUG nova.compute.manager [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 550.409943] env[62965]: ERROR nova.compute.manager [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 52a9bae7-4ed9-425c-908c-82c0d761c688, please check neutron logs for more information. [ 550.409943] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 550.409943] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 550.409943] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 550.409943] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 550.409943] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 550.409943] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 550.409943] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 550.409943] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 550.409943] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 550.409943] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 550.409943] env[62965]: ERROR nova.compute.manager raise self.value [ 550.409943] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 550.409943] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 550.409943] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 550.409943] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 550.410410] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 550.410410] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 550.410410] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 52a9bae7-4ed9-425c-908c-82c0d761c688, please check neutron logs for more information. [ 550.410410] env[62965]: ERROR nova.compute.manager [ 550.410733] env[62965]: Traceback (most recent call last): [ 550.410733] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 550.410733] env[62965]: listener.cb(fileno) [ 550.410733] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 550.410733] env[62965]: result = function(*args, **kwargs) [ 550.410733] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 550.410733] env[62965]: return func(*args, **kwargs) [ 550.410733] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 550.410733] env[62965]: raise e [ 550.410733] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 550.410733] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 550.410733] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 550.410733] env[62965]: created_port_ids = self._update_ports_for_instance( [ 550.410733] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 550.410733] env[62965]: with excutils.save_and_reraise_exception(): [ 550.410733] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 550.410733] env[62965]: self.force_reraise() [ 550.410733] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 550.410733] env[62965]: raise self.value [ 550.410733] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 550.410733] env[62965]: updated_port = self._update_port( [ 550.410733] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 550.410733] env[62965]: _ensure_no_port_binding_failure(port) [ 550.410733] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 550.410733] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 550.410733] env[62965]: nova.exception.PortBindingFailed: Binding failed for port 52a9bae7-4ed9-425c-908c-82c0d761c688, please check neutron logs for more information. [ 550.410733] env[62965]: Removing descriptor: 16 [ 550.411845] env[62965]: ERROR nova.compute.manager [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 52a9bae7-4ed9-425c-908c-82c0d761c688, please check neutron logs for more information. [ 550.411845] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Traceback (most recent call last): [ 550.411845] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 550.411845] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] yield resources [ 550.411845] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 550.411845] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] self.driver.spawn(context, instance, image_meta, [ 550.411845] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 550.411845] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] self._vmops.spawn(context, instance, image_meta, injected_files, [ 550.411845] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 550.411845] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] vm_ref = self.build_virtual_machine(instance, [ 550.411845] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 550.412148] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] vif_infos = vmwarevif.get_vif_info(self._session, [ 550.412148] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 550.412148] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] for vif in network_info: [ 550.412148] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 550.412148] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] return self._sync_wrapper(fn, *args, **kwargs) [ 550.412148] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 550.412148] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] self.wait() [ 550.412148] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 550.412148] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] self[:] = self._gt.wait() [ 550.412148] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 550.412148] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] return self._exit_event.wait() [ 550.412148] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 550.412148] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] result = hub.switch() [ 550.412578] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 550.412578] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] return self.greenlet.switch() [ 550.412578] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 550.412578] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] result = function(*args, **kwargs) [ 550.412578] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 550.412578] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] return func(*args, **kwargs) [ 550.412578] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 550.412578] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] raise e [ 550.412578] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 550.412578] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] nwinfo = self.network_api.allocate_for_instance( [ 550.412578] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 550.412578] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] created_port_ids = self._update_ports_for_instance( [ 550.412578] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 550.412970] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] with excutils.save_and_reraise_exception(): [ 550.412970] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 550.412970] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] self.force_reraise() [ 550.412970] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 550.412970] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] raise self.value [ 550.412970] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 550.412970] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] updated_port = self._update_port( [ 550.412970] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 550.412970] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] _ensure_no_port_binding_failure(port) [ 550.412970] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 550.412970] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] raise exception.PortBindingFailed(port_id=port['id']) [ 550.412970] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] nova.exception.PortBindingFailed: Binding failed for port 52a9bae7-4ed9-425c-908c-82c0d761c688, please check neutron logs for more information. [ 550.412970] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] [ 550.413427] env[62965]: INFO nova.compute.manager [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Terminating instance [ 550.481846] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Task: {'id': task-1389723, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065994} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 550.482244] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 550.483206] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3b67a1b-3a05-4ce6-9ab9-75ba23f5d84f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.509495] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Reconfiguring VM instance instance-00000001 to attach disk [datastore1] 532ae225-da05-4cbb-9bb4-c72a8451acff/532ae225-da05-4cbb-9bb4-c72a8451acff.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 550.512135] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0105869d-2e47-4bb0-973b-00781cd3166f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.535615] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Waiting for the task: (returnval){ [ 550.535615] env[62965]: value = "task-1389724" [ 550.535615] env[62965]: _type = "Task" [ 550.535615] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 550.547700] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Task: {'id': task-1389724, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 550.646254] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Acquiring lock "refresh_cache-ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 550.646254] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Acquired lock "refresh_cache-ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 550.646393] env[62965]: DEBUG nova.network.neutron [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 550.657745] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df21c1ca-1fc4-47f4-ac57-e13acedaa185 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.673026] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f20a7180-4eab-4f46-89d1-84a8f8fc8d7b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.714238] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ff6584b-a6ba-417d-9dcb-c2f81681f1ae {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.723172] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5680e79b-aaca-4286-b7db-e96b7f985646 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.740426] env[62965]: DEBUG nova.compute.provider_tree [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 550.797902] env[62965]: DEBUG nova.network.neutron [req-6c57f27a-5181-4374-99f4-75aa53d35f2c req-0f826206-4c4d-4c3b-981a-cec9bff14732 service nova] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 550.819459] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] Acquiring lock "d66e557f-fbd5-4369-bdcb-14f25c3bbc7b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.819693] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] Lock "d66e557f-fbd5-4369-bdcb-14f25c3bbc7b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.919176] env[62965]: DEBUG nova.compute.manager [req-fc35d09a-663e-40d1-8c32-3ba79cff769d req-c45a5e33-de2d-4dd4-8d13-b6fa55dfb0af service nova] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Received event network-changed-0a2a87c1-e174-4aa2-b17c-54f557ffd83b {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 550.920355] env[62965]: DEBUG nova.compute.manager [req-fc35d09a-663e-40d1-8c32-3ba79cff769d req-c45a5e33-de2d-4dd4-8d13-b6fa55dfb0af service nova] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Refreshing instance network info cache due to event network-changed-0a2a87c1-e174-4aa2-b17c-54f557ffd83b. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 550.920355] env[62965]: DEBUG oslo_concurrency.lockutils [req-fc35d09a-663e-40d1-8c32-3ba79cff769d req-c45a5e33-de2d-4dd4-8d13-b6fa55dfb0af service nova] Acquiring lock "refresh_cache-ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 550.921198] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] Acquiring lock "refresh_cache-1e128a17-cc1a-4aa0-8020-9c01775d9f75" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 550.921198] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] Acquired lock "refresh_cache-1e128a17-cc1a-4aa0-8020-9c01775d9f75" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 550.921198] env[62965]: DEBUG nova.network.neutron [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 550.987360] env[62965]: DEBUG nova.network.neutron [req-6c57f27a-5181-4374-99f4-75aa53d35f2c req-0f826206-4c4d-4c3b-981a-cec9bff14732 service nova] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.045662] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Task: {'id': task-1389724, 'name': ReconfigVM_Task, 'duration_secs': 0.30334} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 551.045943] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Reconfigured VM instance instance-00000001 to attach disk [datastore1] 532ae225-da05-4cbb-9bb4-c72a8451acff/532ae225-da05-4cbb-9bb4-c72a8451acff.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 551.046634] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e5957eda-2d7d-4f18-af98-87e90908e08e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.053187] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Waiting for the task: (returnval){ [ 551.053187] env[62965]: value = "task-1389725" [ 551.053187] env[62965]: _type = "Task" [ 551.053187] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 551.063185] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Task: {'id': task-1389725, 'name': Rename_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 551.197080] env[62965]: DEBUG nova.network.neutron [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 551.244264] env[62965]: DEBUG nova.scheduler.client.report [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 551.275771] env[62965]: DEBUG oslo_concurrency.lockutils [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquiring lock "2931c8b9-ed5f-4994-a4d5-50e1369d67e9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.276343] env[62965]: DEBUG oslo_concurrency.lockutils [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "2931c8b9-ed5f-4994-a4d5-50e1369d67e9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.307495] env[62965]: DEBUG nova.network.neutron [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Successfully created port: fafa1864-43d6-483c-8a81-484a454f3f86 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 551.350448] env[62965]: DEBUG nova.compute.manager [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 551.390116] env[62965]: DEBUG nova.virt.hardware [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 551.390754] env[62965]: DEBUG nova.virt.hardware [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 551.390754] env[62965]: DEBUG nova.virt.hardware [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 551.390754] env[62965]: DEBUG nova.virt.hardware [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 551.390921] env[62965]: DEBUG nova.virt.hardware [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 551.390957] env[62965]: DEBUG nova.virt.hardware [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 551.391240] env[62965]: DEBUG nova.virt.hardware [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 551.391416] env[62965]: DEBUG nova.virt.hardware [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 551.391580] env[62965]: DEBUG nova.virt.hardware [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 551.392009] env[62965]: DEBUG nova.virt.hardware [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 551.392241] env[62965]: DEBUG nova.virt.hardware [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 551.393147] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d9f96cc-3640-453e-9346-4edc84ba875f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.401887] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-845a2cea-2437-47e0-8133-2597d8b82bea {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.421200] env[62965]: DEBUG nova.network.neutron [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.460252] env[62965]: DEBUG nova.network.neutron [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 551.493369] env[62965]: DEBUG oslo_concurrency.lockutils [req-6c57f27a-5181-4374-99f4-75aa53d35f2c req-0f826206-4c4d-4c3b-981a-cec9bff14732 service nova] Releasing lock "refresh_cache-585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 551.571949] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Task: {'id': task-1389725, 'name': Rename_Task, 'duration_secs': 0.129754} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 551.573177] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 551.573177] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-55b023f9-6df4-4105-ad2b-55a4fecfbc91 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.579697] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Waiting for the task: (returnval){ [ 551.579697] env[62965]: value = "task-1389726" [ 551.579697] env[62965]: _type = "Task" [ 551.579697] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 551.592144] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Task: {'id': task-1389726, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 551.749770] env[62965]: DEBUG oslo_concurrency.lockutils [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.432s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.752938] env[62965]: DEBUG nova.compute.manager [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 551.763837] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.906s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.763837] env[62965]: INFO nova.compute.claims [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 551.772942] env[62965]: DEBUG nova.network.neutron [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.923273] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Releasing lock "refresh_cache-ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 551.924042] env[62965]: DEBUG nova.compute.manager [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 551.924042] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 551.924300] env[62965]: DEBUG oslo_concurrency.lockutils [req-fc35d09a-663e-40d1-8c32-3ba79cff769d req-c45a5e33-de2d-4dd4-8d13-b6fa55dfb0af service nova] Acquired lock "refresh_cache-ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 551.924419] env[62965]: DEBUG nova.network.neutron [req-fc35d09a-663e-40d1-8c32-3ba79cff769d req-c45a5e33-de2d-4dd4-8d13-b6fa55dfb0af service nova] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Refreshing network info cache for port 0a2a87c1-e174-4aa2-b17c-54f557ffd83b {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 551.925568] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-479fde88-86f9-47a5-8989-dcf2d928d0ba {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.939542] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70dff217-b9c9-4310-8a6a-41f831f5f373 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.971327] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb could not be found. [ 551.971600] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 551.971870] env[62965]: INFO nova.compute.manager [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Took 0.05 seconds to destroy the instance on the hypervisor. [ 551.972192] env[62965]: DEBUG oslo.service.loopingcall [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 551.972472] env[62965]: DEBUG nova.compute.manager [-] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 551.973213] env[62965]: DEBUG nova.network.neutron [-] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 552.092848] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Task: {'id': task-1389726, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 552.148724] env[62965]: DEBUG nova.network.neutron [-] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 552.262087] env[62965]: DEBUG nova.compute.utils [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 552.263582] env[62965]: DEBUG nova.compute.manager [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 552.263731] env[62965]: DEBUG nova.network.neutron [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 552.276130] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] Releasing lock "refresh_cache-1e128a17-cc1a-4aa0-8020-9c01775d9f75" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 552.276130] env[62965]: DEBUG nova.compute.manager [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 552.276372] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 552.276533] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cd11d9bf-0936-4070-96e7-6e4435009b93 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.292408] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-486b8d67-78c9-49cd-b5e8-babb29aaf3f6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.314585] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Acquiring lock "f619dd08-0cdf-486f-8f13-707b254a8593" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.314786] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Lock "f619dd08-0cdf-486f-8f13-707b254a8593" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.322705] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1e128a17-cc1a-4aa0-8020-9c01775d9f75 could not be found. [ 552.322705] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 552.322705] env[62965]: INFO nova.compute.manager [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Took 0.05 seconds to destroy the instance on the hypervisor. [ 552.322954] env[62965]: DEBUG oslo.service.loopingcall [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 552.323151] env[62965]: DEBUG nova.compute.manager [-] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 552.323247] env[62965]: DEBUG nova.network.neutron [-] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 552.422869] env[62965]: DEBUG nova.network.neutron [-] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 552.520083] env[62965]: DEBUG nova.policy [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd6897fd3a289468facf6080a29c90dc9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8ddb2d2fb1284e438ade1df789cedf7a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 552.536872] env[62965]: DEBUG nova.network.neutron [req-fc35d09a-663e-40d1-8c32-3ba79cff769d req-c45a5e33-de2d-4dd4-8d13-b6fa55dfb0af service nova] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 552.596270] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Task: {'id': task-1389726, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 552.653919] env[62965]: DEBUG nova.network.neutron [-] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.768220] env[62965]: DEBUG nova.compute.manager [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 552.787985] env[62965]: DEBUG nova.network.neutron [req-fc35d09a-663e-40d1-8c32-3ba79cff769d req-c45a5e33-de2d-4dd4-8d13-b6fa55dfb0af service nova] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.925427] env[62965]: DEBUG nova.network.neutron [-] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.954290] env[62965]: DEBUG nova.compute.manager [req-0df389b0-24f0-4665-9801-9a6d9f752718 req-92fb7bcb-19fd-44a6-822d-48471f67a2aa service nova] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Received event network-vif-deleted-c2c98b22-9f4c-4cdf-9b90-5856a23c4cce {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 552.954645] env[62965]: DEBUG nova.compute.manager [req-0df389b0-24f0-4665-9801-9a6d9f752718 req-92fb7bcb-19fd-44a6-822d-48471f67a2aa service nova] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Received event network-changed-52a9bae7-4ed9-425c-908c-82c0d761c688 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 552.954645] env[62965]: DEBUG nova.compute.manager [req-0df389b0-24f0-4665-9801-9a6d9f752718 req-92fb7bcb-19fd-44a6-822d-48471f67a2aa service nova] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Refreshing instance network info cache due to event network-changed-52a9bae7-4ed9-425c-908c-82c0d761c688. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 552.955018] env[62965]: DEBUG oslo_concurrency.lockutils [req-0df389b0-24f0-4665-9801-9a6d9f752718 req-92fb7bcb-19fd-44a6-822d-48471f67a2aa service nova] Acquiring lock "refresh_cache-1e128a17-cc1a-4aa0-8020-9c01775d9f75" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 552.955527] env[62965]: DEBUG oslo_concurrency.lockutils [req-0df389b0-24f0-4665-9801-9a6d9f752718 req-92fb7bcb-19fd-44a6-822d-48471f67a2aa service nova] Acquired lock "refresh_cache-1e128a17-cc1a-4aa0-8020-9c01775d9f75" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 552.955527] env[62965]: DEBUG nova.network.neutron [req-0df389b0-24f0-4665-9801-9a6d9f752718 req-92fb7bcb-19fd-44a6-822d-48471f67a2aa service nova] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Refreshing network info cache for port 52a9bae7-4ed9-425c-908c-82c0d761c688 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 553.056901] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c7f5264-db6c-4dea-a9fb-395d4433aa3c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.067094] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-304cda09-e91a-4a5d-971c-feee4802b7e3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.108409] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50fe2daf-5895-4fff-8067-6bf797f556e5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.116576] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Task: {'id': task-1389726, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 553.121956] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3efe886c-f857-4ca7-b125-1af3589f5fb9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.137477] env[62965]: DEBUG nova.compute.provider_tree [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 553.156419] env[62965]: INFO nova.compute.manager [-] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Took 1.18 seconds to deallocate network for instance. [ 553.160491] env[62965]: DEBUG nova.compute.claims [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 553.160491] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.292382] env[62965]: DEBUG oslo_concurrency.lockutils [req-fc35d09a-663e-40d1-8c32-3ba79cff769d req-c45a5e33-de2d-4dd4-8d13-b6fa55dfb0af service nova] Releasing lock "refresh_cache-ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 553.351047] env[62965]: DEBUG nova.compute.manager [req-1e4c8924-3f87-4441-9b2d-a771443facb1 req-5e519480-b535-4d6e-927c-b280b02f75c4 service nova] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Received event network-vif-deleted-0a2a87c1-e174-4aa2-b17c-54f557ffd83b {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 553.431977] env[62965]: INFO nova.compute.manager [-] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Took 1.11 seconds to deallocate network for instance. [ 553.435697] env[62965]: DEBUG nova.compute.claims [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 553.436058] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.505778] env[62965]: DEBUG nova.network.neutron [req-0df389b0-24f0-4665-9801-9a6d9f752718 req-92fb7bcb-19fd-44a6-822d-48471f67a2aa service nova] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 553.614064] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Task: {'id': task-1389726, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 553.641420] env[62965]: DEBUG nova.scheduler.client.report [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 553.790722] env[62965]: DEBUG nova.compute.manager [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 553.824100] env[62965]: DEBUG nova.virt.hardware [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 553.824100] env[62965]: DEBUG nova.virt.hardware [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 553.824100] env[62965]: DEBUG nova.virt.hardware [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 553.824469] env[62965]: DEBUG nova.virt.hardware [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 553.824469] env[62965]: DEBUG nova.virt.hardware [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 553.824469] env[62965]: DEBUG nova.virt.hardware [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 553.824469] env[62965]: DEBUG nova.virt.hardware [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 553.824626] env[62965]: DEBUG nova.virt.hardware [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 553.824858] env[62965]: DEBUG nova.virt.hardware [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 553.824944] env[62965]: DEBUG nova.virt.hardware [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 553.825072] env[62965]: DEBUG nova.virt.hardware [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 553.826295] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e42282d8-25a0-4a66-9ff6-1e555627bc0c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.838077] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a375f03-0a52-48be-a8c2-fa279ce8abf2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.027152] env[62965]: DEBUG nova.network.neutron [req-0df389b0-24f0-4665-9801-9a6d9f752718 req-92fb7bcb-19fd-44a6-822d-48471f67a2aa service nova] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 554.115753] env[62965]: DEBUG oslo_vmware.api [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Task: {'id': task-1389726, 'name': PowerOnVM_Task, 'duration_secs': 2.267156} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 554.116568] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 554.116568] env[62965]: INFO nova.compute.manager [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Took 12.51 seconds to spawn the instance on the hypervisor. [ 554.116568] env[62965]: DEBUG nova.compute.manager [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 554.119248] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fd216e9-70e2-4bbf-bfa7-f92e121baa81 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.148340] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.387s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 554.148803] env[62965]: DEBUG nova.compute.manager [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 554.154722] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.889s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 554.158588] env[62965]: INFO nova.compute.claims [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 554.173670] env[62965]: DEBUG nova.network.neutron [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Successfully created port: 7ca66b62-dbd5-44c0-8bef-3ca7cf104ca1 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 554.529350] env[62965]: DEBUG oslo_concurrency.lockutils [req-0df389b0-24f0-4665-9801-9a6d9f752718 req-92fb7bcb-19fd-44a6-822d-48471f67a2aa service nova] Releasing lock "refresh_cache-1e128a17-cc1a-4aa0-8020-9c01775d9f75" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 554.529621] env[62965]: DEBUG nova.compute.manager [req-0df389b0-24f0-4665-9801-9a6d9f752718 req-92fb7bcb-19fd-44a6-822d-48471f67a2aa service nova] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Received event network-vif-deleted-52a9bae7-4ed9-425c-908c-82c0d761c688 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 554.640024] env[62965]: INFO nova.compute.manager [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Took 17.36 seconds to build instance. [ 554.665539] env[62965]: DEBUG nova.compute.utils [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 554.673139] env[62965]: DEBUG nova.compute.manager [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 554.673139] env[62965]: DEBUG nova.network.neutron [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 554.906517] env[62965]: DEBUG nova.policy [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c5815f8fbefe45668f3cc4ea11fc251a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8d5533b4d28d4a9cb17069e5a95324d0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 555.147022] env[62965]: DEBUG oslo_concurrency.lockutils [None req-93fc9916-b57d-44d3-9601-b219a4677061 tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Lock "532ae225-da05-4cbb-9bb4-c72a8451acff" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.874s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 555.173496] env[62965]: DEBUG nova.compute.manager [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 555.521423] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16c0bab1-494b-4242-80ec-e06a76906b35 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.530489] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d2d41b1-f050-488e-956d-9fb3298dac6c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.561162] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-779eadba-640a-43c7-995d-0f6187527471 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.568480] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e34b553-573f-4c26-8530-c1fcf8604d14 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.583023] env[62965]: DEBUG nova.compute.provider_tree [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 555.652665] env[62965]: DEBUG nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 555.854457] env[62965]: DEBUG nova.compute.manager [None req-9dc0225b-a506-4ca2-ad3b-a152ea2ee2be tempest-ServerDiagnosticsV248Test-714282837 tempest-ServerDiagnosticsV248Test-714282837-project-admin] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 555.854457] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0afd7bcd-dedf-450b-81d0-6b5a22305415 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.865450] env[62965]: INFO nova.compute.manager [None req-9dc0225b-a506-4ca2-ad3b-a152ea2ee2be tempest-ServerDiagnosticsV248Test-714282837 tempest-ServerDiagnosticsV248Test-714282837-project-admin] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Retrieving diagnostics [ 555.866014] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-181cf8ef-73f8-4132-bff2-7c111696bebc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.002272] env[62965]: DEBUG nova.network.neutron [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Successfully created port: 688eba8b-b5ba-4d6b-9baf-2300a4d2dac2 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 556.089297] env[62965]: DEBUG nova.scheduler.client.report [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 556.191158] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.197723] env[62965]: DEBUG nova.compute.manager [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 556.348794] env[62965]: DEBUG nova.virt.hardware [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 556.348794] env[62965]: DEBUG nova.virt.hardware [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 556.348794] env[62965]: DEBUG nova.virt.hardware [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 556.348970] env[62965]: DEBUG nova.virt.hardware [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 556.348970] env[62965]: DEBUG nova.virt.hardware [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 556.349039] env[62965]: DEBUG nova.virt.hardware [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 556.349567] env[62965]: DEBUG nova.virt.hardware [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 556.349567] env[62965]: DEBUG nova.virt.hardware [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 556.349567] env[62965]: DEBUG nova.virt.hardware [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 556.352344] env[62965]: DEBUG nova.virt.hardware [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 556.352344] env[62965]: DEBUG nova.virt.hardware [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 556.352344] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa6b08d3-04cc-46f0-8512-2f2cf4b867d1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.365972] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fb50fee-d8af-4de1-a3ea-550ffc7b5250 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.496302] env[62965]: ERROR nova.compute.manager [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fafa1864-43d6-483c-8a81-484a454f3f86, please check neutron logs for more information. [ 556.496302] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 556.496302] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 556.496302] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 556.496302] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 556.496302] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 556.496302] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 556.496302] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 556.496302] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.496302] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 556.496302] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.496302] env[62965]: ERROR nova.compute.manager raise self.value [ 556.496302] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 556.496302] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 556.496302] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.496302] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 556.497126] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.497126] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 556.497126] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fafa1864-43d6-483c-8a81-484a454f3f86, please check neutron logs for more information. [ 556.497126] env[62965]: ERROR nova.compute.manager [ 556.497126] env[62965]: Traceback (most recent call last): [ 556.497126] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 556.497126] env[62965]: listener.cb(fileno) [ 556.497126] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 556.497126] env[62965]: result = function(*args, **kwargs) [ 556.497126] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 556.497126] env[62965]: return func(*args, **kwargs) [ 556.497126] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 556.497126] env[62965]: raise e [ 556.497126] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 556.497126] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 556.497126] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 556.497126] env[62965]: created_port_ids = self._update_ports_for_instance( [ 556.497126] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 556.497126] env[62965]: with excutils.save_and_reraise_exception(): [ 556.497126] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.497126] env[62965]: self.force_reraise() [ 556.497126] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.497126] env[62965]: raise self.value [ 556.497126] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 556.497126] env[62965]: updated_port = self._update_port( [ 556.497126] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.497126] env[62965]: _ensure_no_port_binding_failure(port) [ 556.497126] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.497126] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 556.499483] env[62965]: nova.exception.PortBindingFailed: Binding failed for port fafa1864-43d6-483c-8a81-484a454f3f86, please check neutron logs for more information. [ 556.499483] env[62965]: Removing descriptor: 18 [ 556.499483] env[62965]: ERROR nova.compute.manager [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fafa1864-43d6-483c-8a81-484a454f3f86, please check neutron logs for more information. [ 556.499483] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Traceback (most recent call last): [ 556.499483] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 556.499483] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] yield resources [ 556.499483] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 556.499483] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] self.driver.spawn(context, instance, image_meta, [ 556.499483] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 556.499483] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 556.499483] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 556.499483] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] vm_ref = self.build_virtual_machine(instance, [ 556.499788] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 556.499788] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] vif_infos = vmwarevif.get_vif_info(self._session, [ 556.499788] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 556.499788] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] for vif in network_info: [ 556.499788] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 556.499788] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] return self._sync_wrapper(fn, *args, **kwargs) [ 556.499788] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 556.499788] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] self.wait() [ 556.499788] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 556.499788] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] self[:] = self._gt.wait() [ 556.499788] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 556.499788] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] return self._exit_event.wait() [ 556.499788] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 556.500529] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] result = hub.switch() [ 556.500529] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 556.500529] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] return self.greenlet.switch() [ 556.500529] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 556.500529] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] result = function(*args, **kwargs) [ 556.500529] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 556.500529] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] return func(*args, **kwargs) [ 556.500529] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 556.500529] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] raise e [ 556.500529] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 556.500529] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] nwinfo = self.network_api.allocate_for_instance( [ 556.500529] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 556.500529] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] created_port_ids = self._update_ports_for_instance( [ 556.501595] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 556.501595] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] with excutils.save_and_reraise_exception(): [ 556.501595] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.501595] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] self.force_reraise() [ 556.501595] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.501595] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] raise self.value [ 556.501595] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 556.501595] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] updated_port = self._update_port( [ 556.501595] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.501595] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] _ensure_no_port_binding_failure(port) [ 556.501595] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.501595] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] raise exception.PortBindingFailed(port_id=port['id']) [ 556.502012] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] nova.exception.PortBindingFailed: Binding failed for port fafa1864-43d6-483c-8a81-484a454f3f86, please check neutron logs for more information. [ 556.502012] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] [ 556.502012] env[62965]: INFO nova.compute.manager [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Terminating instance [ 556.597967] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.445s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 556.599155] env[62965]: DEBUG nova.compute.manager [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 556.603058] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 10.468s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.603208] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 556.603738] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62965) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 556.605389] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.923s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.607351] env[62965]: INFO nova.compute.claims [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 556.611928] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad102f17-948d-4d9e-9eab-3cfa0df5d31e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.618546] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3e958c7-9709-4f7a-acfb-c982627d2b5a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.636359] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bd0dd4b-1e34-4a27-af78-9fd1eaf0a36f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.644406] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30ae16b2-d27a-4985-b4bd-385e415ac285 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.682267] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181438MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62965) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 556.685314] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.929202] env[62965]: DEBUG nova.compute.manager [req-d943dbbb-cf24-491f-b68e-fcc715267a44 req-d5be98e9-3d21-444f-9f89-27a775f026a8 service nova] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Received event network-changed-fafa1864-43d6-483c-8a81-484a454f3f86 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 556.929202] env[62965]: DEBUG nova.compute.manager [req-d943dbbb-cf24-491f-b68e-fcc715267a44 req-d5be98e9-3d21-444f-9f89-27a775f026a8 service nova] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Refreshing instance network info cache due to event network-changed-fafa1864-43d6-483c-8a81-484a454f3f86. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 556.929202] env[62965]: DEBUG oslo_concurrency.lockutils [req-d943dbbb-cf24-491f-b68e-fcc715267a44 req-d5be98e9-3d21-444f-9f89-27a775f026a8 service nova] Acquiring lock "refresh_cache-c46f89de-5156-49b4-b8f7-f60d78f1134b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 556.929202] env[62965]: DEBUG oslo_concurrency.lockutils [req-d943dbbb-cf24-491f-b68e-fcc715267a44 req-d5be98e9-3d21-444f-9f89-27a775f026a8 service nova] Acquired lock "refresh_cache-c46f89de-5156-49b4-b8f7-f60d78f1134b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 556.929202] env[62965]: DEBUG nova.network.neutron [req-d943dbbb-cf24-491f-b68e-fcc715267a44 req-d5be98e9-3d21-444f-9f89-27a775f026a8 service nova] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Refreshing network info cache for port fafa1864-43d6-483c-8a81-484a454f3f86 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 557.003607] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Acquiring lock "refresh_cache-c46f89de-5156-49b4-b8f7-f60d78f1134b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 557.111918] env[62965]: DEBUG nova.compute.utils [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 557.114992] env[62965]: DEBUG nova.compute.manager [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 557.114992] env[62965]: DEBUG nova.network.neutron [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 557.226029] env[62965]: DEBUG nova.policy [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a6788bd87356434cae74fe73701dd01e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'be2499f8fb014b9384c22dae0b09ed6b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 557.495893] env[62965]: DEBUG nova.network.neutron [req-d943dbbb-cf24-491f-b68e-fcc715267a44 req-d5be98e9-3d21-444f-9f89-27a775f026a8 service nova] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 557.619153] env[62965]: DEBUG nova.compute.manager [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 557.839683] env[62965]: DEBUG nova.network.neutron [req-d943dbbb-cf24-491f-b68e-fcc715267a44 req-d5be98e9-3d21-444f-9f89-27a775f026a8 service nova] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.882807] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b06ad13-b071-4e56-822e-2921ecd1efac {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.891415] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a04aee1-39e9-4c48-b1c9-14f4351d71df {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.924060] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caa5381c-0dbc-40fa-a1ac-d955f105f004 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.931992] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9947ddd0-9732-42be-95ea-5e17412302f9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.946520] env[62965]: DEBUG nova.compute.provider_tree [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 558.178507] env[62965]: DEBUG nova.network.neutron [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Successfully created port: d43a2c0d-b31a-4ee0-bccb-821a706f2c80 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 558.304180] env[62965]: ERROR nova.compute.manager [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7ca66b62-dbd5-44c0-8bef-3ca7cf104ca1, please check neutron logs for more information. [ 558.304180] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 558.304180] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 558.304180] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 558.304180] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 558.304180] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 558.304180] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 558.304180] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 558.304180] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 558.304180] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 558.304180] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 558.304180] env[62965]: ERROR nova.compute.manager raise self.value [ 558.304180] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 558.304180] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 558.304180] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 558.304180] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 558.304933] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 558.304933] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 558.304933] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7ca66b62-dbd5-44c0-8bef-3ca7cf104ca1, please check neutron logs for more information. [ 558.304933] env[62965]: ERROR nova.compute.manager [ 558.304933] env[62965]: Traceback (most recent call last): [ 558.304933] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 558.304933] env[62965]: listener.cb(fileno) [ 558.304933] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 558.304933] env[62965]: result = function(*args, **kwargs) [ 558.304933] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 558.304933] env[62965]: return func(*args, **kwargs) [ 558.304933] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 558.304933] env[62965]: raise e [ 558.304933] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 558.304933] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 558.304933] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 558.304933] env[62965]: created_port_ids = self._update_ports_for_instance( [ 558.304933] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 558.304933] env[62965]: with excutils.save_and_reraise_exception(): [ 558.304933] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 558.304933] env[62965]: self.force_reraise() [ 558.304933] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 558.304933] env[62965]: raise self.value [ 558.304933] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 558.304933] env[62965]: updated_port = self._update_port( [ 558.304933] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 558.304933] env[62965]: _ensure_no_port_binding_failure(port) [ 558.304933] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 558.304933] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 558.307266] env[62965]: nova.exception.PortBindingFailed: Binding failed for port 7ca66b62-dbd5-44c0-8bef-3ca7cf104ca1, please check neutron logs for more information. [ 558.307266] env[62965]: Removing descriptor: 16 [ 558.307266] env[62965]: ERROR nova.compute.manager [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7ca66b62-dbd5-44c0-8bef-3ca7cf104ca1, please check neutron logs for more information. [ 558.307266] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Traceback (most recent call last): [ 558.307266] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 558.307266] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] yield resources [ 558.307266] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 558.307266] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] self.driver.spawn(context, instance, image_meta, [ 558.307266] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 558.307266] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 558.307266] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 558.307266] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] vm_ref = self.build_virtual_machine(instance, [ 558.307750] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 558.307750] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] vif_infos = vmwarevif.get_vif_info(self._session, [ 558.307750] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 558.307750] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] for vif in network_info: [ 558.307750] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 558.307750] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] return self._sync_wrapper(fn, *args, **kwargs) [ 558.307750] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 558.307750] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] self.wait() [ 558.307750] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 558.307750] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] self[:] = self._gt.wait() [ 558.307750] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 558.307750] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] return self._exit_event.wait() [ 558.307750] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 558.308187] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] result = hub.switch() [ 558.308187] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 558.308187] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] return self.greenlet.switch() [ 558.308187] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 558.308187] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] result = function(*args, **kwargs) [ 558.308187] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 558.308187] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] return func(*args, **kwargs) [ 558.308187] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 558.308187] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] raise e [ 558.308187] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 558.308187] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] nwinfo = self.network_api.allocate_for_instance( [ 558.308187] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 558.308187] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] created_port_ids = self._update_ports_for_instance( [ 558.308575] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 558.308575] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] with excutils.save_and_reraise_exception(): [ 558.308575] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 558.308575] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] self.force_reraise() [ 558.308575] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 558.308575] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] raise self.value [ 558.308575] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 558.308575] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] updated_port = self._update_port( [ 558.308575] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 558.308575] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] _ensure_no_port_binding_failure(port) [ 558.308575] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 558.308575] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] raise exception.PortBindingFailed(port_id=port['id']) [ 558.308935] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] nova.exception.PortBindingFailed: Binding failed for port 7ca66b62-dbd5-44c0-8bef-3ca7cf104ca1, please check neutron logs for more information. [ 558.308935] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] [ 558.308935] env[62965]: INFO nova.compute.manager [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Terminating instance [ 558.343455] env[62965]: DEBUG oslo_concurrency.lockutils [req-d943dbbb-cf24-491f-b68e-fcc715267a44 req-d5be98e9-3d21-444f-9f89-27a775f026a8 service nova] Releasing lock "refresh_cache-c46f89de-5156-49b4-b8f7-f60d78f1134b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 558.343597] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Acquired lock "refresh_cache-c46f89de-5156-49b4-b8f7-f60d78f1134b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 558.343781] env[62965]: DEBUG nova.network.neutron [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 558.452490] env[62965]: DEBUG nova.scheduler.client.report [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 558.637904] env[62965]: DEBUG nova.compute.manager [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 558.663350] env[62965]: DEBUG nova.virt.hardware [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 558.663510] env[62965]: DEBUG nova.virt.hardware [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 558.663678] env[62965]: DEBUG nova.virt.hardware [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 558.663920] env[62965]: DEBUG nova.virt.hardware [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 558.664156] env[62965]: DEBUG nova.virt.hardware [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 558.664349] env[62965]: DEBUG nova.virt.hardware [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 558.664583] env[62965]: DEBUG nova.virt.hardware [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 558.665184] env[62965]: DEBUG nova.virt.hardware [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 558.665184] env[62965]: DEBUG nova.virt.hardware [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 558.665184] env[62965]: DEBUG nova.virt.hardware [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 558.665902] env[62965]: DEBUG nova.virt.hardware [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 558.666587] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed1f6f07-92b0-4edf-a4ea-54e736365a37 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.676378] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4265e23-eb7b-42a3-b81c-02f88b0d0c2b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.815117] env[62965]: DEBUG oslo_concurrency.lockutils [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] Acquiring lock "refresh_cache-709bac33-9605-4254-bf38-49fc7892b15e" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 558.815275] env[62965]: DEBUG oslo_concurrency.lockutils [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] Acquired lock "refresh_cache-709bac33-9605-4254-bf38-49fc7892b15e" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 558.815388] env[62965]: DEBUG nova.network.neutron [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 558.902249] env[62965]: DEBUG nova.network.neutron [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 558.938454] env[62965]: DEBUG oslo_concurrency.lockutils [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] Acquiring lock "61702d54-59d2-4d13-9eed-59b052b7fe74" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.938638] env[62965]: DEBUG oslo_concurrency.lockutils [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] Lock "61702d54-59d2-4d13-9eed-59b052b7fe74" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.960171] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.355s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 558.960624] env[62965]: DEBUG nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 558.964093] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.210s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.965891] env[62965]: INFO nova.compute.claims [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 559.407412] env[62965]: DEBUG nova.network.neutron [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 559.421273] env[62965]: DEBUG nova.network.neutron [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 559.466155] env[62965]: DEBUG nova.compute.utils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 559.467550] env[62965]: DEBUG nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 559.467720] env[62965]: DEBUG nova.network.neutron [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 559.686479] env[62965]: DEBUG nova.network.neutron [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 559.702849] env[62965]: DEBUG nova.policy [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '75d7e9144a57445eba2b2e6384b0c14e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '948074c42c564d59b7e0895f3b3c62cd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 559.910365] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Releasing lock "refresh_cache-c46f89de-5156-49b4-b8f7-f60d78f1134b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.910666] env[62965]: DEBUG nova.compute.manager [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 559.911332] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 559.911332] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3d245a02-ea1f-4f8e-a07a-bf7b0dd1b494 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.926641] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0315a779-5927-4f31-828e-44ac62d6e77a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.952020] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c46f89de-5156-49b4-b8f7-f60d78f1134b could not be found. [ 559.952020] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 559.955207] env[62965]: INFO nova.compute.manager [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 559.955207] env[62965]: DEBUG oslo.service.loopingcall [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 559.955207] env[62965]: DEBUG nova.compute.manager [-] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 559.955207] env[62965]: DEBUG nova.network.neutron [-] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 559.971466] env[62965]: DEBUG nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 560.013674] env[62965]: DEBUG nova.network.neutron [-] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 560.188830] env[62965]: DEBUG oslo_concurrency.lockutils [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] Releasing lock "refresh_cache-709bac33-9605-4254-bf38-49fc7892b15e" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 560.189375] env[62965]: DEBUG nova.compute.manager [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 560.189614] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 560.190354] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7c844279-9775-4e12-a067-76c88deb47dc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.205899] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-025a9ecc-fd9f-4578-a117-5c712bd31d37 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.236057] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 709bac33-9605-4254-bf38-49fc7892b15e could not be found. [ 560.236304] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 560.236484] env[62965]: INFO nova.compute.manager [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Took 0.05 seconds to destroy the instance on the hypervisor. [ 560.237179] env[62965]: DEBUG oslo.service.loopingcall [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 560.239426] env[62965]: DEBUG nova.compute.manager [-] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 560.239613] env[62965]: DEBUG nova.network.neutron [-] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 560.293780] env[62965]: DEBUG nova.network.neutron [-] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 560.336367] env[62965]: DEBUG oslo_concurrency.lockutils [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] Acquiring lock "59187ca2-0482-4603-8689-542d2bf5d023" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.336367] env[62965]: DEBUG oslo_concurrency.lockutils [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] Lock "59187ca2-0482-4603-8689-542d2bf5d023" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.354358] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c4929e3-e62e-40ef-b249-fef56bed36de {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.364296] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13fc796c-08f1-4f79-826a-5a436da67e64 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.409590] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-389cf6ca-01fb-4d6a-a5a8-e119777ba8c2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.417338] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40c4dbfd-f7d6-4e67-adef-0fc31802edd6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.426187] env[62965]: DEBUG oslo_concurrency.lockutils [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] Acquiring lock "53519d3e-34d2-410f-ac19-69a015e407b9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.426187] env[62965]: DEBUG oslo_concurrency.lockutils [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] Lock "53519d3e-34d2-410f-ac19-69a015e407b9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.437798] env[62965]: DEBUG nova.compute.provider_tree [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 560.515767] env[62965]: DEBUG nova.network.neutron [-] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.798640] env[62965]: DEBUG nova.network.neutron [-] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.804945] env[62965]: DEBUG nova.compute.manager [req-df75dc78-af68-4c19-8f30-adfde2f9a2d9 req-f9fc394f-88c7-457c-8404-ca84949283aa service nova] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Received event network-vif-deleted-fafa1864-43d6-483c-8a81-484a454f3f86 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 560.805912] env[62965]: DEBUG nova.compute.manager [req-df75dc78-af68-4c19-8f30-adfde2f9a2d9 req-f9fc394f-88c7-457c-8404-ca84949283aa service nova] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Received event network-changed-7ca66b62-dbd5-44c0-8bef-3ca7cf104ca1 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 560.805912] env[62965]: DEBUG nova.compute.manager [req-df75dc78-af68-4c19-8f30-adfde2f9a2d9 req-f9fc394f-88c7-457c-8404-ca84949283aa service nova] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Refreshing instance network info cache due to event network-changed-7ca66b62-dbd5-44c0-8bef-3ca7cf104ca1. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 560.805912] env[62965]: DEBUG oslo_concurrency.lockutils [req-df75dc78-af68-4c19-8f30-adfde2f9a2d9 req-f9fc394f-88c7-457c-8404-ca84949283aa service nova] Acquiring lock "refresh_cache-709bac33-9605-4254-bf38-49fc7892b15e" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 560.805912] env[62965]: DEBUG oslo_concurrency.lockutils [req-df75dc78-af68-4c19-8f30-adfde2f9a2d9 req-f9fc394f-88c7-457c-8404-ca84949283aa service nova] Acquired lock "refresh_cache-709bac33-9605-4254-bf38-49fc7892b15e" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 560.805912] env[62965]: DEBUG nova.network.neutron [req-df75dc78-af68-4c19-8f30-adfde2f9a2d9 req-f9fc394f-88c7-457c-8404-ca84949283aa service nova] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Refreshing network info cache for port 7ca66b62-dbd5-44c0-8bef-3ca7cf104ca1 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 560.944351] env[62965]: DEBUG nova.network.neutron [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Successfully created port: d49a2434-0352-4b6f-95a8-3dbf7c5ef1cd {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 560.948370] env[62965]: DEBUG nova.scheduler.client.report [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 560.984605] env[62965]: DEBUG nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 561.017600] env[62965]: INFO nova.compute.manager [-] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Took 1.06 seconds to deallocate network for instance. [ 561.020096] env[62965]: DEBUG nova.compute.claims [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 561.020271] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.024990] env[62965]: DEBUG nova.virt.hardware [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 561.025231] env[62965]: DEBUG nova.virt.hardware [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 561.025379] env[62965]: DEBUG nova.virt.hardware [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 561.025551] env[62965]: DEBUG nova.virt.hardware [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 561.025762] env[62965]: DEBUG nova.virt.hardware [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 561.025969] env[62965]: DEBUG nova.virt.hardware [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 561.026195] env[62965]: DEBUG nova.virt.hardware [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 561.026348] env[62965]: DEBUG nova.virt.hardware [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 561.026508] env[62965]: DEBUG nova.virt.hardware [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 561.026659] env[62965]: DEBUG nova.virt.hardware [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 561.026827] env[62965]: DEBUG nova.virt.hardware [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 561.027759] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-655824b5-f650-4eb4-ab57-bdcd4abe19bf {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.037142] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e35f105-5701-4e4f-a537-2512d253cc4b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.235198] env[62965]: ERROR nova.compute.manager [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 688eba8b-b5ba-4d6b-9baf-2300a4d2dac2, please check neutron logs for more information. [ 561.235198] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 561.235198] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 561.235198] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 561.235198] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 561.235198] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 561.235198] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 561.235198] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 561.235198] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 561.235198] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 561.235198] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 561.235198] env[62965]: ERROR nova.compute.manager raise self.value [ 561.235198] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 561.235198] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 561.235198] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 561.235198] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 561.236109] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 561.236109] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 561.236109] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 688eba8b-b5ba-4d6b-9baf-2300a4d2dac2, please check neutron logs for more information. [ 561.236109] env[62965]: ERROR nova.compute.manager [ 561.236109] env[62965]: Traceback (most recent call last): [ 561.236109] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 561.236109] env[62965]: listener.cb(fileno) [ 561.236109] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 561.236109] env[62965]: result = function(*args, **kwargs) [ 561.236109] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 561.236109] env[62965]: return func(*args, **kwargs) [ 561.236109] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 561.236109] env[62965]: raise e [ 561.236109] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 561.236109] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 561.236109] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 561.236109] env[62965]: created_port_ids = self._update_ports_for_instance( [ 561.236109] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 561.236109] env[62965]: with excutils.save_and_reraise_exception(): [ 561.236109] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 561.236109] env[62965]: self.force_reraise() [ 561.236109] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 561.236109] env[62965]: raise self.value [ 561.236109] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 561.236109] env[62965]: updated_port = self._update_port( [ 561.236109] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 561.236109] env[62965]: _ensure_no_port_binding_failure(port) [ 561.236109] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 561.236109] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 561.237904] env[62965]: nova.exception.PortBindingFailed: Binding failed for port 688eba8b-b5ba-4d6b-9baf-2300a4d2dac2, please check neutron logs for more information. [ 561.237904] env[62965]: Removing descriptor: 17 [ 561.237904] env[62965]: ERROR nova.compute.manager [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 688eba8b-b5ba-4d6b-9baf-2300a4d2dac2, please check neutron logs for more information. [ 561.237904] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Traceback (most recent call last): [ 561.237904] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 561.237904] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] yield resources [ 561.237904] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 561.237904] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] self.driver.spawn(context, instance, image_meta, [ 561.237904] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 561.237904] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] self._vmops.spawn(context, instance, image_meta, injected_files, [ 561.237904] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 561.237904] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] vm_ref = self.build_virtual_machine(instance, [ 561.240134] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 561.240134] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] vif_infos = vmwarevif.get_vif_info(self._session, [ 561.240134] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 561.240134] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] for vif in network_info: [ 561.240134] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 561.240134] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] return self._sync_wrapper(fn, *args, **kwargs) [ 561.240134] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 561.240134] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] self.wait() [ 561.240134] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 561.240134] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] self[:] = self._gt.wait() [ 561.240134] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 561.240134] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] return self._exit_event.wait() [ 561.240134] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 561.240612] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] result = hub.switch() [ 561.240612] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 561.240612] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] return self.greenlet.switch() [ 561.240612] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 561.240612] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] result = function(*args, **kwargs) [ 561.240612] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 561.240612] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] return func(*args, **kwargs) [ 561.240612] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 561.240612] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] raise e [ 561.240612] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 561.240612] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] nwinfo = self.network_api.allocate_for_instance( [ 561.240612] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 561.240612] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] created_port_ids = self._update_ports_for_instance( [ 561.240979] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 561.240979] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] with excutils.save_and_reraise_exception(): [ 561.240979] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 561.240979] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] self.force_reraise() [ 561.240979] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 561.240979] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] raise self.value [ 561.240979] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 561.240979] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] updated_port = self._update_port( [ 561.240979] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 561.240979] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] _ensure_no_port_binding_failure(port) [ 561.240979] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 561.240979] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] raise exception.PortBindingFailed(port_id=port['id']) [ 561.241360] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] nova.exception.PortBindingFailed: Binding failed for port 688eba8b-b5ba-4d6b-9baf-2300a4d2dac2, please check neutron logs for more information. [ 561.241360] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] [ 561.241360] env[62965]: INFO nova.compute.manager [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Terminating instance [ 561.302296] env[62965]: INFO nova.compute.manager [-] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Took 1.06 seconds to deallocate network for instance. [ 561.304788] env[62965]: DEBUG nova.compute.claims [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 561.304965] env[62965]: DEBUG oslo_concurrency.lockutils [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.351093] env[62965]: DEBUG nova.network.neutron [req-df75dc78-af68-4c19-8f30-adfde2f9a2d9 req-f9fc394f-88c7-457c-8404-ca84949283aa service nova] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 561.453961] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.490s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 561.456133] env[62965]: DEBUG nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 561.463641] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.435s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.470089] env[62965]: DEBUG nova.network.neutron [req-df75dc78-af68-4c19-8f30-adfde2f9a2d9 req-f9fc394f-88c7-457c-8404-ca84949283aa service nova] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.496544] env[62965]: DEBUG nova.compute.manager [req-85b6c9b6-e7c6-454a-b7e8-e64f38b4bce4 req-dc8dd6ea-e71f-43a8-8716-199ded1eadd9 service nova] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Received event network-changed-688eba8b-b5ba-4d6b-9baf-2300a4d2dac2 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 561.496873] env[62965]: DEBUG nova.compute.manager [req-85b6c9b6-e7c6-454a-b7e8-e64f38b4bce4 req-dc8dd6ea-e71f-43a8-8716-199ded1eadd9 service nova] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Refreshing instance network info cache due to event network-changed-688eba8b-b5ba-4d6b-9baf-2300a4d2dac2. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 561.497028] env[62965]: DEBUG oslo_concurrency.lockutils [req-85b6c9b6-e7c6-454a-b7e8-e64f38b4bce4 req-dc8dd6ea-e71f-43a8-8716-199ded1eadd9 service nova] Acquiring lock "refresh_cache-7780b346-28a4-4d80-9f83-a452c23741ec" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 561.497219] env[62965]: DEBUG oslo_concurrency.lockutils [req-85b6c9b6-e7c6-454a-b7e8-e64f38b4bce4 req-dc8dd6ea-e71f-43a8-8716-199ded1eadd9 service nova] Acquired lock "refresh_cache-7780b346-28a4-4d80-9f83-a452c23741ec" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 561.498026] env[62965]: DEBUG nova.network.neutron [req-85b6c9b6-e7c6-454a-b7e8-e64f38b4bce4 req-dc8dd6ea-e71f-43a8-8716-199ded1eadd9 service nova] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Refreshing network info cache for port 688eba8b-b5ba-4d6b-9baf-2300a4d2dac2 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 561.740313] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] Acquiring lock "refresh_cache-7780b346-28a4-4d80-9f83-a452c23741ec" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 561.962708] env[62965]: DEBUG nova.compute.utils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 561.964165] env[62965]: DEBUG nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 561.964906] env[62965]: DEBUG nova.network.neutron [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 561.973102] env[62965]: DEBUG oslo_concurrency.lockutils [req-df75dc78-af68-4c19-8f30-adfde2f9a2d9 req-f9fc394f-88c7-457c-8404-ca84949283aa service nova] Releasing lock "refresh_cache-709bac33-9605-4254-bf38-49fc7892b15e" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 561.973382] env[62965]: DEBUG nova.compute.manager [req-df75dc78-af68-4c19-8f30-adfde2f9a2d9 req-f9fc394f-88c7-457c-8404-ca84949283aa service nova] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Received event network-vif-deleted-7ca66b62-dbd5-44c0-8bef-3ca7cf104ca1 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 562.046592] env[62965]: DEBUG nova.network.neutron [req-85b6c9b6-e7c6-454a-b7e8-e64f38b4bce4 req-dc8dd6ea-e71f-43a8-8716-199ded1eadd9 service nova] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 562.121482] env[62965]: DEBUG nova.policy [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '75d7e9144a57445eba2b2e6384b0c14e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '948074c42c564d59b7e0895f3b3c62cd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 562.326937] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d655449-a2c7-4eac-96e7-a98efc491bb4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.335941] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5b5557d-6869-4cd6-a273-c7c3dbbf2295 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.368426] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42fe891b-2515-44c5-9c40-3506c47b6fa2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.373779] env[62965]: DEBUG nova.network.neutron [req-85b6c9b6-e7c6-454a-b7e8-e64f38b4bce4 req-dc8dd6ea-e71f-43a8-8716-199ded1eadd9 service nova] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.378320] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-618c30ca-daf1-42bb-886d-0f68b3664500 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.392579] env[62965]: DEBUG nova.compute.provider_tree [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 562.467844] env[62965]: DEBUG nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 562.625979] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] Acquiring lock "f594c50e-25b3-4e85-840c-7dfe0994a2c2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.628028] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] Lock "f594c50e-25b3-4e85-840c-7dfe0994a2c2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.833208] env[62965]: ERROR nova.compute.manager [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d43a2c0d-b31a-4ee0-bccb-821a706f2c80, please check neutron logs for more information. [ 562.833208] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 562.833208] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 562.833208] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 562.833208] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 562.833208] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 562.833208] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 562.833208] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 562.833208] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 562.833208] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 562.833208] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 562.833208] env[62965]: ERROR nova.compute.manager raise self.value [ 562.833208] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 562.833208] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 562.833208] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 562.833208] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 562.833944] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 562.833944] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 562.833944] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d43a2c0d-b31a-4ee0-bccb-821a706f2c80, please check neutron logs for more information. [ 562.833944] env[62965]: ERROR nova.compute.manager [ 562.833944] env[62965]: Traceback (most recent call last): [ 562.833944] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 562.833944] env[62965]: listener.cb(fileno) [ 562.833944] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 562.833944] env[62965]: result = function(*args, **kwargs) [ 562.833944] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 562.833944] env[62965]: return func(*args, **kwargs) [ 562.833944] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 562.833944] env[62965]: raise e [ 562.833944] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 562.833944] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 562.833944] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 562.833944] env[62965]: created_port_ids = self._update_ports_for_instance( [ 562.833944] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 562.833944] env[62965]: with excutils.save_and_reraise_exception(): [ 562.833944] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 562.833944] env[62965]: self.force_reraise() [ 562.833944] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 562.833944] env[62965]: raise self.value [ 562.833944] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 562.833944] env[62965]: updated_port = self._update_port( [ 562.833944] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 562.833944] env[62965]: _ensure_no_port_binding_failure(port) [ 562.833944] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 562.833944] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 562.834789] env[62965]: nova.exception.PortBindingFailed: Binding failed for port d43a2c0d-b31a-4ee0-bccb-821a706f2c80, please check neutron logs for more information. [ 562.834789] env[62965]: Removing descriptor: 18 [ 562.834789] env[62965]: ERROR nova.compute.manager [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d43a2c0d-b31a-4ee0-bccb-821a706f2c80, please check neutron logs for more information. [ 562.834789] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Traceback (most recent call last): [ 562.834789] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 562.834789] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] yield resources [ 562.834789] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 562.834789] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] self.driver.spawn(context, instance, image_meta, [ 562.834789] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 562.834789] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] self._vmops.spawn(context, instance, image_meta, injected_files, [ 562.834789] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 562.834789] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] vm_ref = self.build_virtual_machine(instance, [ 562.835189] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 562.835189] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] vif_infos = vmwarevif.get_vif_info(self._session, [ 562.835189] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 562.835189] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] for vif in network_info: [ 562.835189] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 562.835189] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] return self._sync_wrapper(fn, *args, **kwargs) [ 562.835189] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 562.835189] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] self.wait() [ 562.835189] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 562.835189] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] self[:] = self._gt.wait() [ 562.835189] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 562.835189] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] return self._exit_event.wait() [ 562.835189] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 562.835544] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] result = hub.switch() [ 562.835544] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 562.835544] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] return self.greenlet.switch() [ 562.835544] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 562.835544] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] result = function(*args, **kwargs) [ 562.835544] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 562.835544] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] return func(*args, **kwargs) [ 562.835544] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 562.835544] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] raise e [ 562.835544] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 562.835544] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] nwinfo = self.network_api.allocate_for_instance( [ 562.835544] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 562.835544] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] created_port_ids = self._update_ports_for_instance( [ 562.836728] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 562.836728] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] with excutils.save_and_reraise_exception(): [ 562.836728] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 562.836728] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] self.force_reraise() [ 562.836728] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 562.836728] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] raise self.value [ 562.836728] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 562.836728] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] updated_port = self._update_port( [ 562.836728] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 562.836728] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] _ensure_no_port_binding_failure(port) [ 562.836728] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 562.836728] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] raise exception.PortBindingFailed(port_id=port['id']) [ 562.837263] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] nova.exception.PortBindingFailed: Binding failed for port d43a2c0d-b31a-4ee0-bccb-821a706f2c80, please check neutron logs for more information. [ 562.837263] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] [ 562.837263] env[62965]: INFO nova.compute.manager [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Terminating instance [ 562.878398] env[62965]: DEBUG oslo_concurrency.lockutils [req-85b6c9b6-e7c6-454a-b7e8-e64f38b4bce4 req-dc8dd6ea-e71f-43a8-8716-199ded1eadd9 service nova] Releasing lock "refresh_cache-7780b346-28a4-4d80-9f83-a452c23741ec" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 562.878828] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] Acquired lock "refresh_cache-7780b346-28a4-4d80-9f83-a452c23741ec" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 562.879172] env[62965]: DEBUG nova.network.neutron [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 562.896753] env[62965]: DEBUG nova.scheduler.client.report [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 563.213589] env[62965]: DEBUG nova.network.neutron [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Successfully created port: 3c249e6a-17cb-4734-b532-233e86c66d5c {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 563.343795] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Acquiring lock "refresh_cache-fa55e796-81d0-46b7-b025-9119bae2f305" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 563.343973] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Acquired lock "refresh_cache-fa55e796-81d0-46b7-b025-9119bae2f305" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 563.344161] env[62965]: DEBUG nova.network.neutron [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 563.377701] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Acquiring lock "09b62788-3082-4c99-9dfc-0843c4988feb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.377701] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Lock "09b62788-3082-4c99-9dfc-0843c4988feb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.400785] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.941s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 563.401461] env[62965]: ERROR nova.compute.manager [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c2c98b22-9f4c-4cdf-9b90-5856a23c4cce, please check neutron logs for more information. [ 563.401461] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Traceback (most recent call last): [ 563.401461] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 563.401461] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] self.driver.spawn(context, instance, image_meta, [ 563.401461] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 563.401461] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 563.401461] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 563.401461] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] vm_ref = self.build_virtual_machine(instance, [ 563.401461] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 563.401461] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] vif_infos = vmwarevif.get_vif_info(self._session, [ 563.401461] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 563.401968] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] for vif in network_info: [ 563.401968] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 563.401968] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] return self._sync_wrapper(fn, *args, **kwargs) [ 563.401968] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 563.401968] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] self.wait() [ 563.401968] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 563.401968] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] self[:] = self._gt.wait() [ 563.401968] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 563.401968] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] return self._exit_event.wait() [ 563.401968] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 563.401968] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] result = hub.switch() [ 563.401968] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 563.401968] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] return self.greenlet.switch() [ 563.402417] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 563.402417] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] result = function(*args, **kwargs) [ 563.402417] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 563.402417] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] return func(*args, **kwargs) [ 563.402417] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 563.402417] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] raise e [ 563.402417] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 563.402417] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] nwinfo = self.network_api.allocate_for_instance( [ 563.402417] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 563.402417] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] created_port_ids = self._update_ports_for_instance( [ 563.402417] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 563.402417] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] with excutils.save_and_reraise_exception(): [ 563.402417] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 563.402906] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] self.force_reraise() [ 563.402906] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 563.402906] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] raise self.value [ 563.402906] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 563.402906] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] updated_port = self._update_port( [ 563.402906] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 563.402906] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] _ensure_no_port_binding_failure(port) [ 563.402906] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 563.402906] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] raise exception.PortBindingFailed(port_id=port['id']) [ 563.402906] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] nova.exception.PortBindingFailed: Binding failed for port c2c98b22-9f4c-4cdf-9b90-5856a23c4cce, please check neutron logs for more information. [ 563.402906] env[62965]: ERROR nova.compute.manager [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] [ 563.403264] env[62965]: DEBUG nova.compute.utils [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Binding failed for port c2c98b22-9f4c-4cdf-9b90-5856a23c4cce, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 563.403264] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.243s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.412261] env[62965]: DEBUG nova.compute.manager [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Build of instance 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5 was re-scheduled: Binding failed for port c2c98b22-9f4c-4cdf-9b90-5856a23c4cce, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 563.412261] env[62965]: DEBUG nova.compute.manager [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 563.413540] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] Acquiring lock "refresh_cache-585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 563.413540] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] Acquired lock "refresh_cache-585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 563.413540] env[62965]: DEBUG nova.network.neutron [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 563.418380] env[62965]: DEBUG nova.network.neutron [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 563.481745] env[62965]: DEBUG nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 563.523955] env[62965]: DEBUG nova.virt.hardware [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 563.524219] env[62965]: DEBUG nova.virt.hardware [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 563.524428] env[62965]: DEBUG nova.virt.hardware [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 563.524627] env[62965]: DEBUG nova.virt.hardware [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 563.524751] env[62965]: DEBUG nova.virt.hardware [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 563.525036] env[62965]: DEBUG nova.virt.hardware [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 563.525119] env[62965]: DEBUG nova.virt.hardware [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 563.525265] env[62965]: DEBUG nova.virt.hardware [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 563.526294] env[62965]: DEBUG nova.virt.hardware [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 563.526294] env[62965]: DEBUG nova.virt.hardware [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 563.526294] env[62965]: DEBUG nova.virt.hardware [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 563.527060] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-240457d6-9e0b-4392-9559-6a8cc89d9b4f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.535984] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bbfe10e-5bc2-4983-8f96-688f4696b414 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.589208] env[62965]: DEBUG nova.network.neutron [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 563.872917] env[62965]: DEBUG nova.compute.manager [req-83df38a6-b74d-4241-a883-508c7cae3b06 req-b068c2c0-f7b4-4e14-8475-e1cef0792abb service nova] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Received event network-changed-d43a2c0d-b31a-4ee0-bccb-821a706f2c80 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 563.873206] env[62965]: DEBUG nova.compute.manager [req-83df38a6-b74d-4241-a883-508c7cae3b06 req-b068c2c0-f7b4-4e14-8475-e1cef0792abb service nova] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Refreshing instance network info cache due to event network-changed-d43a2c0d-b31a-4ee0-bccb-821a706f2c80. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 563.873404] env[62965]: DEBUG oslo_concurrency.lockutils [req-83df38a6-b74d-4241-a883-508c7cae3b06 req-b068c2c0-f7b4-4e14-8475-e1cef0792abb service nova] Acquiring lock "refresh_cache-fa55e796-81d0-46b7-b025-9119bae2f305" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 563.888033] env[62965]: DEBUG nova.network.neutron [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 563.917968] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] Acquiring lock "63c54ed4-4525-42f6-a99f-033f7b9838a7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.918249] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] Lock "63c54ed4-4525-42f6-a99f-033f7b9838a7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.958377] env[62965]: DEBUG nova.network.neutron [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 564.097934] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] Releasing lock "refresh_cache-7780b346-28a4-4d80-9f83-a452c23741ec" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 564.097934] env[62965]: DEBUG nova.compute.manager [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 564.097934] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 564.102021] env[62965]: ERROR nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d49a2434-0352-4b6f-95a8-3dbf7c5ef1cd, please check neutron logs for more information. [ 564.102021] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 564.102021] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 564.102021] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 564.102021] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 564.102021] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 564.102021] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 564.102021] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 564.102021] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 564.102021] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 564.102021] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 564.102021] env[62965]: ERROR nova.compute.manager raise self.value [ 564.102021] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 564.102021] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 564.102021] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 564.102021] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 564.102642] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 564.102642] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 564.102642] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d49a2434-0352-4b6f-95a8-3dbf7c5ef1cd, please check neutron logs for more information. [ 564.102642] env[62965]: ERROR nova.compute.manager [ 564.102642] env[62965]: Traceback (most recent call last): [ 564.102642] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 564.102642] env[62965]: listener.cb(fileno) [ 564.102642] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 564.102642] env[62965]: result = function(*args, **kwargs) [ 564.102642] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 564.102642] env[62965]: return func(*args, **kwargs) [ 564.102642] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 564.102642] env[62965]: raise e [ 564.102642] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 564.102642] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 564.102642] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 564.102642] env[62965]: created_port_ids = self._update_ports_for_instance( [ 564.102642] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 564.102642] env[62965]: with excutils.save_and_reraise_exception(): [ 564.102642] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 564.102642] env[62965]: self.force_reraise() [ 564.102642] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 564.102642] env[62965]: raise self.value [ 564.102642] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 564.102642] env[62965]: updated_port = self._update_port( [ 564.102642] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 564.102642] env[62965]: _ensure_no_port_binding_failure(port) [ 564.102642] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 564.102642] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 564.103394] env[62965]: nova.exception.PortBindingFailed: Binding failed for port d49a2434-0352-4b6f-95a8-3dbf7c5ef1cd, please check neutron logs for more information. [ 564.103394] env[62965]: Removing descriptor: 16 [ 564.103394] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-33e36c73-ac83-48c6-bc75-eb2a4dbfc8d3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.103394] env[62965]: ERROR nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d49a2434-0352-4b6f-95a8-3dbf7c5ef1cd, please check neutron logs for more information. [ 564.103394] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Traceback (most recent call last): [ 564.103394] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 564.103394] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] yield resources [ 564.103394] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 564.103394] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] self.driver.spawn(context, instance, image_meta, [ 564.103394] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 564.103394] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 564.103872] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 564.103872] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] vm_ref = self.build_virtual_machine(instance, [ 564.103872] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 564.103872] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] vif_infos = vmwarevif.get_vif_info(self._session, [ 564.103872] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 564.103872] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] for vif in network_info: [ 564.103872] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 564.103872] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] return self._sync_wrapper(fn, *args, **kwargs) [ 564.103872] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 564.103872] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] self.wait() [ 564.103872] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 564.103872] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] self[:] = self._gt.wait() [ 564.103872] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 564.104253] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] return self._exit_event.wait() [ 564.104253] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 564.104253] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] result = hub.switch() [ 564.104253] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 564.104253] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] return self.greenlet.switch() [ 564.104253] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 564.104253] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] result = function(*args, **kwargs) [ 564.104253] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 564.104253] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] return func(*args, **kwargs) [ 564.104253] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 564.104253] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] raise e [ 564.104253] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 564.104253] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] nwinfo = self.network_api.allocate_for_instance( [ 564.105327] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 564.105327] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] created_port_ids = self._update_ports_for_instance( [ 564.105327] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 564.105327] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] with excutils.save_and_reraise_exception(): [ 564.105327] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 564.105327] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] self.force_reraise() [ 564.105327] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 564.105327] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] raise self.value [ 564.105327] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 564.105327] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] updated_port = self._update_port( [ 564.105327] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 564.105327] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] _ensure_no_port_binding_failure(port) [ 564.105327] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 564.105660] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] raise exception.PortBindingFailed(port_id=port['id']) [ 564.105660] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] nova.exception.PortBindingFailed: Binding failed for port d49a2434-0352-4b6f-95a8-3dbf7c5ef1cd, please check neutron logs for more information. [ 564.105660] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] [ 564.105660] env[62965]: INFO nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Terminating instance [ 564.115299] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff5d6916-659f-4034-9851-aad9794278c7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.129139] env[62965]: DEBUG nova.network.neutron [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.135042] env[62965]: DEBUG nova.network.neutron [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.153387] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7780b346-28a4-4d80-9f83-a452c23741ec could not be found. [ 564.153650] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 564.153830] env[62965]: INFO nova.compute.manager [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Took 0.06 seconds to destroy the instance on the hypervisor. [ 564.154397] env[62965]: DEBUG oslo.service.loopingcall [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 564.157781] env[62965]: DEBUG nova.compute.manager [-] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 564.157781] env[62965]: DEBUG nova.network.neutron [-] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 564.182796] env[62965]: DEBUG nova.network.neutron [-] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 564.323235] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abded568-8d94-46e4-9495-b11d152117e6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.331871] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3a05119-d5fa-4020-aef9-c50ebf3e3e96 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.365979] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37999bdb-3d0e-435d-90c1-7c0174218f2d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.373613] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1757a6c-a54c-4353-9ffb-a23ae689e46e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.388317] env[62965]: DEBUG nova.compute.provider_tree [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 564.609955] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Acquiring lock "refresh_cache-b811920a-14a2-4c11-8b13-78d920f08a0e" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 564.610177] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Acquired lock "refresh_cache-b811920a-14a2-4c11-8b13-78d920f08a0e" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 564.610324] env[62965]: DEBUG nova.network.neutron [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 564.635718] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] Releasing lock "refresh_cache-585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 564.635849] env[62965]: DEBUG nova.compute.manager [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 564.636017] env[62965]: DEBUG nova.compute.manager [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 564.636159] env[62965]: DEBUG nova.network.neutron [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 564.641247] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Releasing lock "refresh_cache-fa55e796-81d0-46b7-b025-9119bae2f305" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 564.641326] env[62965]: DEBUG nova.compute.manager [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 564.641756] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 564.642213] env[62965]: DEBUG oslo_concurrency.lockutils [req-83df38a6-b74d-4241-a883-508c7cae3b06 req-b068c2c0-f7b4-4e14-8475-e1cef0792abb service nova] Acquired lock "refresh_cache-fa55e796-81d0-46b7-b025-9119bae2f305" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 564.642406] env[62965]: DEBUG nova.network.neutron [req-83df38a6-b74d-4241-a883-508c7cae3b06 req-b068c2c0-f7b4-4e14-8475-e1cef0792abb service nova] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Refreshing network info cache for port d43a2c0d-b31a-4ee0-bccb-821a706f2c80 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 564.643591] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-58442b32-6532-4f5a-bfc3-3d3f9736efc8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.655824] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85024e54-cfdb-4bbc-be5c-6d948f369d45 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.668015] env[62965]: DEBUG nova.network.neutron [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 564.683010] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fa55e796-81d0-46b7-b025-9119bae2f305 could not be found. [ 564.683285] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 564.683555] env[62965]: INFO nova.compute.manager [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Took 0.04 seconds to destroy the instance on the hypervisor. [ 564.683825] env[62965]: DEBUG oslo.service.loopingcall [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 564.684621] env[62965]: DEBUG nova.compute.manager [-] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 564.684762] env[62965]: DEBUG nova.network.neutron [-] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 564.686700] env[62965]: DEBUG nova.network.neutron [-] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.708421] env[62965]: DEBUG nova.network.neutron [-] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 564.834132] env[62965]: DEBUG nova.compute.manager [req-4d2d58c7-beeb-4368-a78e-be8c7dc2c9be req-42e3b8e6-8ed0-4d98-9aef-58f373aa8485 service nova] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Received event network-vif-deleted-688eba8b-b5ba-4d6b-9baf-2300a4d2dac2 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 564.892354] env[62965]: DEBUG nova.scheduler.client.report [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 565.146084] env[62965]: DEBUG nova.network.neutron [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 565.175295] env[62965]: DEBUG nova.network.neutron [req-83df38a6-b74d-4241-a883-508c7cae3b06 req-b068c2c0-f7b4-4e14-8475-e1cef0792abb service nova] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 565.180478] env[62965]: DEBUG nova.network.neutron [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 565.191593] env[62965]: INFO nova.compute.manager [-] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Took 1.03 seconds to deallocate network for instance. [ 565.195168] env[62965]: DEBUG nova.compute.claims [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 565.195168] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.211392] env[62965]: DEBUG nova.network.neutron [-] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 565.351491] env[62965]: DEBUG oslo_concurrency.lockutils [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] Acquiring lock "222f547e-8fce-4576-bafc-809ea2ee8631" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.352150] env[62965]: DEBUG oslo_concurrency.lockutils [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] Lock "222f547e-8fce-4576-bafc-809ea2ee8631" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.399685] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.996s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 565.400338] env[62965]: ERROR nova.compute.manager [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0a2a87c1-e174-4aa2-b17c-54f557ffd83b, please check neutron logs for more information. [ 565.400338] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Traceback (most recent call last): [ 565.400338] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 565.400338] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] self.driver.spawn(context, instance, image_meta, [ 565.400338] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 565.400338] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 565.400338] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 565.400338] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] vm_ref = self.build_virtual_machine(instance, [ 565.400338] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 565.400338] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] vif_infos = vmwarevif.get_vif_info(self._session, [ 565.400338] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 565.400687] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] for vif in network_info: [ 565.400687] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 565.400687] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] return self._sync_wrapper(fn, *args, **kwargs) [ 565.400687] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 565.400687] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] self.wait() [ 565.400687] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 565.400687] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] self[:] = self._gt.wait() [ 565.400687] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 565.400687] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] return self._exit_event.wait() [ 565.400687] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 565.400687] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] result = hub.switch() [ 565.400687] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 565.400687] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] return self.greenlet.switch() [ 565.401052] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 565.401052] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] result = function(*args, **kwargs) [ 565.401052] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 565.401052] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] return func(*args, **kwargs) [ 565.401052] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 565.401052] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] raise e [ 565.401052] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 565.401052] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] nwinfo = self.network_api.allocate_for_instance( [ 565.401052] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 565.401052] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] created_port_ids = self._update_ports_for_instance( [ 565.401052] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 565.401052] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] with excutils.save_and_reraise_exception(): [ 565.401052] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.401399] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] self.force_reraise() [ 565.401399] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.401399] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] raise self.value [ 565.401399] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 565.401399] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] updated_port = self._update_port( [ 565.401399] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.401399] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] _ensure_no_port_binding_failure(port) [ 565.401399] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.401399] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] raise exception.PortBindingFailed(port_id=port['id']) [ 565.401399] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] nova.exception.PortBindingFailed: Binding failed for port 0a2a87c1-e174-4aa2-b17c-54f557ffd83b, please check neutron logs for more information. [ 565.401399] env[62965]: ERROR nova.compute.manager [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] [ 565.401717] env[62965]: DEBUG nova.compute.utils [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Binding failed for port 0a2a87c1-e174-4aa2-b17c-54f557ffd83b, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 565.402262] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.966s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.406198] env[62965]: DEBUG nova.compute.manager [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Build of instance ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb was re-scheduled: Binding failed for port 0a2a87c1-e174-4aa2-b17c-54f557ffd83b, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 565.406641] env[62965]: DEBUG nova.compute.manager [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 565.406872] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Acquiring lock "refresh_cache-ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 565.407061] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Acquired lock "refresh_cache-ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 565.407221] env[62965]: DEBUG nova.network.neutron [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 565.420278] env[62965]: DEBUG nova.network.neutron [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 565.440658] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Acquiring lock "af29af38-bd31-446a-a944-2bb43d9aa6c6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.443505] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Lock "af29af38-bd31-446a-a944-2bb43d9aa6c6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.456131] env[62965]: ERROR nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3c249e6a-17cb-4734-b532-233e86c66d5c, please check neutron logs for more information. [ 565.456131] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 565.456131] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 565.456131] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 565.456131] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 565.456131] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 565.456131] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 565.456131] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 565.456131] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.456131] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 565.456131] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.456131] env[62965]: ERROR nova.compute.manager raise self.value [ 565.456131] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 565.456131] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 565.456131] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.456131] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 565.456615] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.456615] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 565.456615] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3c249e6a-17cb-4734-b532-233e86c66d5c, please check neutron logs for more information. [ 565.456615] env[62965]: ERROR nova.compute.manager [ 565.456615] env[62965]: Traceback (most recent call last): [ 565.456615] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 565.456615] env[62965]: listener.cb(fileno) [ 565.456615] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 565.456615] env[62965]: result = function(*args, **kwargs) [ 565.456615] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 565.456615] env[62965]: return func(*args, **kwargs) [ 565.456615] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 565.456615] env[62965]: raise e [ 565.456615] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 565.456615] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 565.456615] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 565.456615] env[62965]: created_port_ids = self._update_ports_for_instance( [ 565.456615] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 565.456615] env[62965]: with excutils.save_and_reraise_exception(): [ 565.456615] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.456615] env[62965]: self.force_reraise() [ 565.456615] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.456615] env[62965]: raise self.value [ 565.456615] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 565.456615] env[62965]: updated_port = self._update_port( [ 565.456615] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.456615] env[62965]: _ensure_no_port_binding_failure(port) [ 565.456615] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.456615] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 565.457400] env[62965]: nova.exception.PortBindingFailed: Binding failed for port 3c249e6a-17cb-4734-b532-233e86c66d5c, please check neutron logs for more information. [ 565.457400] env[62965]: Removing descriptor: 17 [ 565.457400] env[62965]: ERROR nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3c249e6a-17cb-4734-b532-233e86c66d5c, please check neutron logs for more information. [ 565.457400] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Traceback (most recent call last): [ 565.457400] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 565.457400] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] yield resources [ 565.457400] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 565.457400] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] self.driver.spawn(context, instance, image_meta, [ 565.457400] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 565.457400] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 565.457400] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 565.457400] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] vm_ref = self.build_virtual_machine(instance, [ 565.457804] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 565.457804] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] vif_infos = vmwarevif.get_vif_info(self._session, [ 565.457804] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 565.457804] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] for vif in network_info: [ 565.457804] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 565.457804] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] return self._sync_wrapper(fn, *args, **kwargs) [ 565.457804] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 565.457804] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] self.wait() [ 565.457804] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 565.457804] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] self[:] = self._gt.wait() [ 565.457804] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 565.457804] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] return self._exit_event.wait() [ 565.457804] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 565.458177] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] result = hub.switch() [ 565.458177] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 565.458177] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] return self.greenlet.switch() [ 565.458177] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 565.458177] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] result = function(*args, **kwargs) [ 565.458177] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 565.458177] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] return func(*args, **kwargs) [ 565.458177] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 565.458177] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] raise e [ 565.458177] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 565.458177] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] nwinfo = self.network_api.allocate_for_instance( [ 565.458177] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 565.458177] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] created_port_ids = self._update_ports_for_instance( [ 565.458525] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 565.458525] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] with excutils.save_and_reraise_exception(): [ 565.458525] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.458525] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] self.force_reraise() [ 565.458525] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.458525] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] raise self.value [ 565.458525] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 565.458525] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] updated_port = self._update_port( [ 565.458525] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.458525] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] _ensure_no_port_binding_failure(port) [ 565.458525] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.458525] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] raise exception.PortBindingFailed(port_id=port['id']) [ 565.459419] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] nova.exception.PortBindingFailed: Binding failed for port 3c249e6a-17cb-4734-b532-233e86c66d5c, please check neutron logs for more information. [ 565.459419] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] [ 565.459419] env[62965]: INFO nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Terminating instance [ 565.475377] env[62965]: DEBUG nova.network.neutron [req-83df38a6-b74d-4241-a883-508c7cae3b06 req-b068c2c0-f7b4-4e14-8475-e1cef0792abb service nova] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 565.683451] env[62965]: INFO nova.compute.manager [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] [instance: 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5] Took 1.05 seconds to deallocate network for instance. [ 565.714309] env[62965]: INFO nova.compute.manager [-] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Took 1.03 seconds to deallocate network for instance. [ 565.719289] env[62965]: DEBUG nova.compute.claims [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 565.719739] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.923626] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Releasing lock "refresh_cache-b811920a-14a2-4c11-8b13-78d920f08a0e" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 565.923626] env[62965]: DEBUG nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 565.923626] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 565.923626] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-04ebdcee-eb14-4f2c-805a-c11020205a47 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.933704] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba39b0c2-5146-489e-a22e-15da305f6270 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.971899] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Acquiring lock "refresh_cache-4a75ddc4-add2-45dc-b086-bc163990db4f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 565.971899] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Acquired lock "refresh_cache-4a75ddc4-add2-45dc-b086-bc163990db4f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 565.972130] env[62965]: DEBUG nova.network.neutron [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 565.973759] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b811920a-14a2-4c11-8b13-78d920f08a0e could not be found. [ 565.974593] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 565.974593] env[62965]: INFO nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Took 0.05 seconds to destroy the instance on the hypervisor. [ 565.974593] env[62965]: DEBUG oslo.service.loopingcall [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 565.974913] env[62965]: DEBUG nova.compute.manager [-] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 565.975069] env[62965]: DEBUG nova.network.neutron [-] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 565.977180] env[62965]: DEBUG oslo_concurrency.lockutils [req-83df38a6-b74d-4241-a883-508c7cae3b06 req-b068c2c0-f7b4-4e14-8475-e1cef0792abb service nova] Releasing lock "refresh_cache-fa55e796-81d0-46b7-b025-9119bae2f305" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 565.977379] env[62965]: DEBUG nova.compute.manager [req-83df38a6-b74d-4241-a883-508c7cae3b06 req-b068c2c0-f7b4-4e14-8475-e1cef0792abb service nova] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Received event network-vif-deleted-d43a2c0d-b31a-4ee0-bccb-821a706f2c80 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 566.011016] env[62965]: DEBUG nova.network.neutron [-] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 566.099362] env[62965]: DEBUG oslo_concurrency.lockutils [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Acquiring lock "8a893e6e-37d4-4fd4-b0db-d6633117bdce" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.099362] env[62965]: DEBUG oslo_concurrency.lockutils [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Lock "8a893e6e-37d4-4fd4-b0db-d6633117bdce" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.101393] env[62965]: DEBUG nova.network.neutron [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 566.199126] env[62965]: DEBUG nova.network.neutron [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.446539] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b092d52-816c-401e-9886-3c98c646ca08 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.457078] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bedb536-952d-40bc-95e8-f72975efd05a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.494967] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d673757-9d94-4bdc-98a7-49e8ff363645 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.502635] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ae890fa-3511-4304-8937-c34d35644e84 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.522105] env[62965]: DEBUG nova.network.neutron [-] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.524417] env[62965]: DEBUG nova.compute.provider_tree [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 566.537414] env[62965]: DEBUG nova.network.neutron [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 566.705476] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Releasing lock "refresh_cache-ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 566.705581] env[62965]: DEBUG nova.compute.manager [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 566.707814] env[62965]: DEBUG nova.compute.manager [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 566.707814] env[62965]: DEBUG nova.network.neutron [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 566.727183] env[62965]: INFO nova.scheduler.client.report [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] Deleted allocations for instance 585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5 [ 566.743434] env[62965]: DEBUG nova.network.neutron [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 566.755611] env[62965]: DEBUG nova.network.neutron [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.922946] env[62965]: DEBUG nova.compute.manager [req-baa8f7cb-9d71-4cfa-b24f-1055faa05d38 req-5b5c0416-95b1-44b3-9157-923089bef278 service nova] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Received event network-changed-d49a2434-0352-4b6f-95a8-3dbf7c5ef1cd {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 566.923159] env[62965]: DEBUG nova.compute.manager [req-baa8f7cb-9d71-4cfa-b24f-1055faa05d38 req-5b5c0416-95b1-44b3-9157-923089bef278 service nova] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Refreshing instance network info cache due to event network-changed-d49a2434-0352-4b6f-95a8-3dbf7c5ef1cd. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 566.923774] env[62965]: DEBUG oslo_concurrency.lockutils [req-baa8f7cb-9d71-4cfa-b24f-1055faa05d38 req-5b5c0416-95b1-44b3-9157-923089bef278 service nova] Acquiring lock "refresh_cache-b811920a-14a2-4c11-8b13-78d920f08a0e" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 566.923774] env[62965]: DEBUG oslo_concurrency.lockutils [req-baa8f7cb-9d71-4cfa-b24f-1055faa05d38 req-5b5c0416-95b1-44b3-9157-923089bef278 service nova] Acquired lock "refresh_cache-b811920a-14a2-4c11-8b13-78d920f08a0e" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 566.923774] env[62965]: DEBUG nova.network.neutron [req-baa8f7cb-9d71-4cfa-b24f-1055faa05d38 req-5b5c0416-95b1-44b3-9157-923089bef278 service nova] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Refreshing network info cache for port d49a2434-0352-4b6f-95a8-3dbf7c5ef1cd {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 566.945852] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] Acquiring lock "8ac4264d-7d75-4c38-b8da-e3d90ffe9363" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.945852] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] Lock "8ac4264d-7d75-4c38-b8da-e3d90ffe9363" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.025329] env[62965]: INFO nova.compute.manager [-] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Took 1.05 seconds to deallocate network for instance. [ 567.029537] env[62965]: DEBUG nova.compute.claims [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 567.029886] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.030628] env[62965]: DEBUG nova.scheduler.client.report [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 567.194989] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Acquiring lock "971f72f8-2686-4ced-a2a1-b7ab8a1ea024" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.195273] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Lock "971f72f8-2686-4ced-a2a1-b7ab8a1ea024" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.236715] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8cfc12a7-f4fa-49df-891c-5f0b2f6f4ff8 tempest-ServersAdminNegativeTestJSON-601945676 tempest-ServersAdminNegativeTestJSON-601945676-project-member] Lock "585a0d73-9a2c-4a1e-8e7e-7b4f406a56d5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.432s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 567.247161] env[62965]: DEBUG nova.network.neutron [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.259464] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Releasing lock "refresh_cache-4a75ddc4-add2-45dc-b086-bc163990db4f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 567.261577] env[62965]: DEBUG nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 567.261577] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 567.261577] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fa3d22bd-ef92-4013-b389-56d64d07b884 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.270632] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5df40c29-04e9-45ad-afea-769d99c1145b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.296179] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4a75ddc4-add2-45dc-b086-bc163990db4f could not be found. [ 567.296397] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 567.296791] env[62965]: INFO nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 567.296892] env[62965]: DEBUG oslo.service.loopingcall [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 567.297219] env[62965]: DEBUG nova.compute.manager [-] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 567.297323] env[62965]: DEBUG nova.network.neutron [-] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 567.317920] env[62965]: DEBUG nova.network.neutron [-] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 567.352328] env[62965]: DEBUG nova.compute.manager [None req-e1b054d4-390d-447b-bc26-d75581f6d75f tempest-ServerDiagnosticsV248Test-714282837 tempest-ServerDiagnosticsV248Test-714282837-project-admin] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 567.353558] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45886ff4-a683-4f8c-8bb6-2469fc5f47b7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.364672] env[62965]: INFO nova.compute.manager [None req-e1b054d4-390d-447b-bc26-d75581f6d75f tempest-ServerDiagnosticsV248Test-714282837 tempest-ServerDiagnosticsV248Test-714282837-project-admin] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Retrieving diagnostics [ 567.368166] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9842cdd5-cc8f-40e3-96a8-a652cd744d2a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.455533] env[62965]: DEBUG nova.network.neutron [req-baa8f7cb-9d71-4cfa-b24f-1055faa05d38 req-5b5c0416-95b1-44b3-9157-923089bef278 service nova] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 567.537598] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.135s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 567.538283] env[62965]: ERROR nova.compute.manager [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 52a9bae7-4ed9-425c-908c-82c0d761c688, please check neutron logs for more information. [ 567.538283] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Traceback (most recent call last): [ 567.538283] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 567.538283] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] self.driver.spawn(context, instance, image_meta, [ 567.538283] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 567.538283] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] self._vmops.spawn(context, instance, image_meta, injected_files, [ 567.538283] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 567.538283] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] vm_ref = self.build_virtual_machine(instance, [ 567.538283] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 567.538283] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] vif_infos = vmwarevif.get_vif_info(self._session, [ 567.538283] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 567.538632] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] for vif in network_info: [ 567.538632] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 567.538632] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] return self._sync_wrapper(fn, *args, **kwargs) [ 567.538632] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 567.538632] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] self.wait() [ 567.538632] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 567.538632] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] self[:] = self._gt.wait() [ 567.538632] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 567.538632] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] return self._exit_event.wait() [ 567.538632] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 567.538632] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] result = hub.switch() [ 567.538632] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 567.538632] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] return self.greenlet.switch() [ 567.538957] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 567.538957] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] result = function(*args, **kwargs) [ 567.538957] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 567.538957] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] return func(*args, **kwargs) [ 567.538957] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 567.538957] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] raise e [ 567.538957] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 567.538957] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] nwinfo = self.network_api.allocate_for_instance( [ 567.538957] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 567.538957] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] created_port_ids = self._update_ports_for_instance( [ 567.538957] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 567.538957] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] with excutils.save_and_reraise_exception(): [ 567.538957] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 567.539347] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] self.force_reraise() [ 567.539347] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 567.539347] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] raise self.value [ 567.539347] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 567.539347] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] updated_port = self._update_port( [ 567.539347] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 567.539347] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] _ensure_no_port_binding_failure(port) [ 567.539347] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 567.539347] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] raise exception.PortBindingFailed(port_id=port['id']) [ 567.539347] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] nova.exception.PortBindingFailed: Binding failed for port 52a9bae7-4ed9-425c-908c-82c0d761c688, please check neutron logs for more information. [ 567.539347] env[62965]: ERROR nova.compute.manager [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] [ 567.539650] env[62965]: DEBUG nova.compute.utils [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Binding failed for port 52a9bae7-4ed9-425c-908c-82c0d761c688, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 567.540546] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.349s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.541742] env[62965]: INFO nova.compute.claims [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 567.546970] env[62965]: DEBUG nova.compute.manager [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Build of instance 1e128a17-cc1a-4aa0-8020-9c01775d9f75 was re-scheduled: Binding failed for port 52a9bae7-4ed9-425c-908c-82c0d761c688, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 567.546970] env[62965]: DEBUG nova.compute.manager [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 567.546970] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] Acquiring lock "refresh_cache-1e128a17-cc1a-4aa0-8020-9c01775d9f75" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 567.546970] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] Acquired lock "refresh_cache-1e128a17-cc1a-4aa0-8020-9c01775d9f75" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 567.547229] env[62965]: DEBUG nova.network.neutron [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 567.571167] env[62965]: DEBUG nova.network.neutron [req-baa8f7cb-9d71-4cfa-b24f-1055faa05d38 req-5b5c0416-95b1-44b3-9157-923089bef278 service nova] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.740867] env[62965]: DEBUG nova.compute.manager [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 567.749549] env[62965]: INFO nova.compute.manager [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb] Took 1.04 seconds to deallocate network for instance. [ 567.820393] env[62965]: DEBUG nova.network.neutron [-] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.997727] env[62965]: DEBUG nova.compute.manager [req-6f7406b3-62b5-4647-b0d2-135af9bf6a6d req-cfe69718-f551-4df7-8145-fe9aec95398e service nova] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Received event network-changed-3c249e6a-17cb-4734-b532-233e86c66d5c {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 567.997912] env[62965]: DEBUG nova.compute.manager [req-6f7406b3-62b5-4647-b0d2-135af9bf6a6d req-cfe69718-f551-4df7-8145-fe9aec95398e service nova] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Refreshing instance network info cache due to event network-changed-3c249e6a-17cb-4734-b532-233e86c66d5c. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 567.998146] env[62965]: DEBUG oslo_concurrency.lockutils [req-6f7406b3-62b5-4647-b0d2-135af9bf6a6d req-cfe69718-f551-4df7-8145-fe9aec95398e service nova] Acquiring lock "refresh_cache-4a75ddc4-add2-45dc-b086-bc163990db4f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 567.998288] env[62965]: DEBUG oslo_concurrency.lockutils [req-6f7406b3-62b5-4647-b0d2-135af9bf6a6d req-cfe69718-f551-4df7-8145-fe9aec95398e service nova] Acquired lock "refresh_cache-4a75ddc4-add2-45dc-b086-bc163990db4f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 567.998449] env[62965]: DEBUG nova.network.neutron [req-6f7406b3-62b5-4647-b0d2-135af9bf6a6d req-cfe69718-f551-4df7-8145-fe9aec95398e service nova] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Refreshing network info cache for port 3c249e6a-17cb-4734-b532-233e86c66d5c {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 568.032607] env[62965]: DEBUG oslo_concurrency.lockutils [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] Acquiring lock "eaf9ca94-4f45-403e-bb06-e9a9adf58f38" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.032895] env[62965]: DEBUG oslo_concurrency.lockutils [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] Lock "eaf9ca94-4f45-403e-bb06-e9a9adf58f38" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.074238] env[62965]: DEBUG oslo_concurrency.lockutils [req-baa8f7cb-9d71-4cfa-b24f-1055faa05d38 req-5b5c0416-95b1-44b3-9157-923089bef278 service nova] Releasing lock "refresh_cache-b811920a-14a2-4c11-8b13-78d920f08a0e" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 568.074495] env[62965]: DEBUG nova.compute.manager [req-baa8f7cb-9d71-4cfa-b24f-1055faa05d38 req-5b5c0416-95b1-44b3-9157-923089bef278 service nova] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Received event network-vif-deleted-d49a2434-0352-4b6f-95a8-3dbf7c5ef1cd {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 568.080414] env[62965]: DEBUG nova.network.neutron [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 568.235124] env[62965]: DEBUG nova.network.neutron [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.292644] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.325669] env[62965]: INFO nova.compute.manager [-] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Took 1.03 seconds to deallocate network for instance. [ 568.328028] env[62965]: DEBUG nova.compute.claims [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 568.328552] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.531747] env[62965]: DEBUG nova.network.neutron [req-6f7406b3-62b5-4647-b0d2-135af9bf6a6d req-cfe69718-f551-4df7-8145-fe9aec95398e service nova] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 568.670340] env[62965]: DEBUG nova.network.neutron [req-6f7406b3-62b5-4647-b0d2-135af9bf6a6d req-cfe69718-f551-4df7-8145-fe9aec95398e service nova] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.737160] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] Releasing lock "refresh_cache-1e128a17-cc1a-4aa0-8020-9c01775d9f75" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 568.737383] env[62965]: DEBUG nova.compute.manager [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 568.737555] env[62965]: DEBUG nova.compute.manager [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 568.737741] env[62965]: DEBUG nova.network.neutron [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 568.780897] env[62965]: DEBUG nova.network.neutron [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 568.795681] env[62965]: INFO nova.scheduler.client.report [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Deleted allocations for instance ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb [ 569.034826] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Acquiring lock "79a4d3c8-ead6-4893-b0a7-89a8ff435e54" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.035113] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Lock "79a4d3c8-ead6-4893-b0a7-89a8ff435e54" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.041070] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb7a7133-d33e-4ecd-9a83-609bf6c58b20 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.049246] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a3cfd78-a26e-4fd0-bcbc-96cb895b3379 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.080242] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dbe27f1-c0ff-4dd9-a6b5-88392bdaac7a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.087685] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78fc76be-7414-4322-b266-8408bce368cd {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.104955] env[62965]: DEBUG nova.compute.provider_tree [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 569.175095] env[62965]: DEBUG oslo_concurrency.lockutils [req-6f7406b3-62b5-4647-b0d2-135af9bf6a6d req-cfe69718-f551-4df7-8145-fe9aec95398e service nova] Releasing lock "refresh_cache-4a75ddc4-add2-45dc-b086-bc163990db4f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 569.176155] env[62965]: DEBUG nova.compute.manager [req-6f7406b3-62b5-4647-b0d2-135af9bf6a6d req-cfe69718-f551-4df7-8145-fe9aec95398e service nova] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Received event network-vif-deleted-3c249e6a-17cb-4734-b532-233e86c66d5c {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 569.286247] env[62965]: DEBUG nova.network.neutron [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.307843] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e30c2aaf-5dcc-4b6a-b8af-4e36ad7d2247 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Lock "ebb7d30d-c9d7-4ed6-8d5b-22c83d7c80bb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.505s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 569.409905] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Acquiring lock "532ae225-da05-4cbb-9bb4-c72a8451acff" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.410187] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Lock "532ae225-da05-4cbb-9bb4-c72a8451acff" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.410392] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Acquiring lock "532ae225-da05-4cbb-9bb4-c72a8451acff-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.410568] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Lock "532ae225-da05-4cbb-9bb4-c72a8451acff-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.410851] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Lock "532ae225-da05-4cbb-9bb4-c72a8451acff-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 569.413158] env[62965]: INFO nova.compute.manager [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Terminating instance [ 569.608918] env[62965]: DEBUG nova.scheduler.client.report [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 569.787997] env[62965]: INFO nova.compute.manager [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] [instance: 1e128a17-cc1a-4aa0-8020-9c01775d9f75] Took 1.05 seconds to deallocate network for instance. [ 569.809940] env[62965]: DEBUG nova.compute.manager [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 569.918785] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Acquiring lock "refresh_cache-532ae225-da05-4cbb-9bb4-c72a8451acff" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 569.918785] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Acquired lock "refresh_cache-532ae225-da05-4cbb-9bb4-c72a8451acff" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 569.918785] env[62965]: DEBUG nova.network.neutron [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 570.120532] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.579s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.120532] env[62965]: DEBUG nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 570.126214] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 13.441s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.345480] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 570.473340] env[62965]: DEBUG nova.network.neutron [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 570.594082] env[62965]: DEBUG nova.network.neutron [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.633956] env[62965]: DEBUG nova.compute.utils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 570.637177] env[62965]: DEBUG nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 570.637373] env[62965]: DEBUG nova.network.neutron [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 570.830796] env[62965]: INFO nova.scheduler.client.report [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] Deleted allocations for instance 1e128a17-cc1a-4aa0-8020-9c01775d9f75 [ 570.903353] env[62965]: DEBUG nova.policy [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '75d7e9144a57445eba2b2e6384b0c14e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '948074c42c564d59b7e0895f3b3c62cd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 571.099709] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Releasing lock "refresh_cache-532ae225-da05-4cbb-9bb4-c72a8451acff" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 571.100206] env[62965]: DEBUG nova.compute.manager [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 571.100300] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 571.101278] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b40700b7-8650-418b-97f2-77f504c28edf {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.111600] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 571.111857] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fbe9bdda-3afd-4e45-af0e-8891ea0af160 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.119532] env[62965]: DEBUG oslo_vmware.api [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Waiting for the task: (returnval){ [ 571.119532] env[62965]: value = "task-1389727" [ 571.119532] env[62965]: _type = "Task" [ 571.119532] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 571.136360] env[62965]: DEBUG oslo_vmware.api [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Task: {'id': task-1389727, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 571.138141] env[62965]: DEBUG nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 571.195228] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 532ae225-da05-4cbb-9bb4-c72a8451acff actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 571.195599] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance c46f89de-5156-49b4-b8f7-f60d78f1134b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 571.195897] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 709bac33-9605-4254-bf38-49fc7892b15e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 571.196091] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 7780b346-28a4-4d80-9f83-a452c23741ec actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 571.196415] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance fa55e796-81d0-46b7-b025-9119bae2f305 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 571.196622] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance b811920a-14a2-4c11-8b13-78d920f08a0e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 571.196772] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 4a75ddc4-add2-45dc-b086-bc163990db4f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 571.196926] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 2870d8cb-8c19-454e-8a6e-1675eba8ddd4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 571.340471] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1f953b02-10c5-45c3-8b2d-76ae2090b430 tempest-FloatingIPsAssociationNegativeTestJSON-552758390 tempest-FloatingIPsAssociationNegativeTestJSON-552758390-project-member] Lock "1e128a17-cc1a-4aa0-8020-9c01775d9f75" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.284s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 571.386859] env[62965]: DEBUG nova.network.neutron [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Successfully created port: 9add2eb2-422d-493e-ac3d-acf980ff6c90 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 571.631614] env[62965]: DEBUG oslo_vmware.api [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Task: {'id': task-1389727, 'name': PowerOffVM_Task, 'duration_secs': 0.125353} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 571.631798] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 571.632274] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 571.632586] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a7617421-d297-4d32-a933-5fbceb8b29d7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.660709] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 571.661208] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Deleting contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 571.661208] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Deleting the datastore file [datastore1] 532ae225-da05-4cbb-9bb4-c72a8451acff {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 571.661640] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4dac388a-f917-445c-87cb-98246072c6d8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.670418] env[62965]: DEBUG oslo_vmware.api [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Waiting for the task: (returnval){ [ 571.670418] env[62965]: value = "task-1389729" [ 571.670418] env[62965]: _type = "Task" [ 571.670418] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 571.681735] env[62965]: DEBUG oslo_vmware.api [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Task: {'id': task-1389729, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 571.704536] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 647a3ba2-8b84-45f9-9d28-130eb0521d1b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 571.843820] env[62965]: DEBUG nova.compute.manager [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 572.150741] env[62965]: DEBUG nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 572.183760] env[62965]: DEBUG oslo_vmware.api [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Task: {'id': task-1389729, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.096365} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 572.186187] env[62965]: DEBUG nova.virt.hardware [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 572.186274] env[62965]: DEBUG nova.virt.hardware [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 572.186685] env[62965]: DEBUG nova.virt.hardware [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 572.186940] env[62965]: DEBUG nova.virt.hardware [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 572.187306] env[62965]: DEBUG nova.virt.hardware [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 572.187488] env[62965]: DEBUG nova.virt.hardware [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 572.187732] env[62965]: DEBUG nova.virt.hardware [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 572.188119] env[62965]: DEBUG nova.virt.hardware [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 572.188378] env[62965]: DEBUG nova.virt.hardware [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 572.188579] env[62965]: DEBUG nova.virt.hardware [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 572.190377] env[62965]: DEBUG nova.virt.hardware [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 572.190377] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 572.190377] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Deleted contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 572.190377] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 572.190377] env[62965]: INFO nova.compute.manager [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Took 1.09 seconds to destroy the instance on the hypervisor. [ 572.190556] env[62965]: DEBUG oslo.service.loopingcall [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 572.190973] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55979c8d-5779-46a4-8524-48ff27842780 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.193930] env[62965]: DEBUG nova.compute.manager [-] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 572.194224] env[62965]: DEBUG nova.network.neutron [-] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 572.200949] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13b43b3d-610f-42fa-a2d6-2d3e0074d896 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.207343] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance d66e557f-fbd5-4369-bdcb-14f25c3bbc7b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 572.237461] env[62965]: DEBUG oslo_concurrency.lockutils [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Acquiring lock "7eee167f-50e3-48d5-9c6d-9115df9bb0b9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.237918] env[62965]: DEBUG oslo_concurrency.lockutils [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Lock "7eee167f-50e3-48d5-9c6d-9115df9bb0b9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.379400] env[62965]: DEBUG oslo_concurrency.lockutils [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.406349] env[62965]: DEBUG nova.network.neutron [-] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 572.596066] env[62965]: ERROR nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9add2eb2-422d-493e-ac3d-acf980ff6c90, please check neutron logs for more information. [ 572.596066] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 572.596066] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 572.596066] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 572.596066] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 572.596066] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 572.596066] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 572.596066] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 572.596066] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.596066] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 572.596066] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.596066] env[62965]: ERROR nova.compute.manager raise self.value [ 572.596066] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 572.596066] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 572.596066] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.596066] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 572.597139] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.597139] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 572.597139] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9add2eb2-422d-493e-ac3d-acf980ff6c90, please check neutron logs for more information. [ 572.597139] env[62965]: ERROR nova.compute.manager [ 572.597139] env[62965]: Traceback (most recent call last): [ 572.597139] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 572.597139] env[62965]: listener.cb(fileno) [ 572.597139] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 572.597139] env[62965]: result = function(*args, **kwargs) [ 572.597139] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 572.597139] env[62965]: return func(*args, **kwargs) [ 572.597139] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 572.597139] env[62965]: raise e [ 572.597139] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 572.597139] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 572.597139] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 572.597139] env[62965]: created_port_ids = self._update_ports_for_instance( [ 572.597139] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 572.597139] env[62965]: with excutils.save_and_reraise_exception(): [ 572.597139] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.597139] env[62965]: self.force_reraise() [ 572.597139] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.597139] env[62965]: raise self.value [ 572.597139] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 572.597139] env[62965]: updated_port = self._update_port( [ 572.597139] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.597139] env[62965]: _ensure_no_port_binding_failure(port) [ 572.597139] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.597139] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 572.598315] env[62965]: nova.exception.PortBindingFailed: Binding failed for port 9add2eb2-422d-493e-ac3d-acf980ff6c90, please check neutron logs for more information. [ 572.598315] env[62965]: Removing descriptor: 19 [ 572.598315] env[62965]: ERROR nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9add2eb2-422d-493e-ac3d-acf980ff6c90, please check neutron logs for more information. [ 572.598315] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Traceback (most recent call last): [ 572.598315] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 572.598315] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] yield resources [ 572.598315] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 572.598315] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] self.driver.spawn(context, instance, image_meta, [ 572.598315] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 572.598315] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 572.598315] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 572.598315] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] vm_ref = self.build_virtual_machine(instance, [ 572.599374] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 572.599374] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] vif_infos = vmwarevif.get_vif_info(self._session, [ 572.599374] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 572.599374] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] for vif in network_info: [ 572.599374] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 572.599374] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] return self._sync_wrapper(fn, *args, **kwargs) [ 572.599374] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 572.599374] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] self.wait() [ 572.599374] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 572.599374] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] self[:] = self._gt.wait() [ 572.599374] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 572.599374] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] return self._exit_event.wait() [ 572.599374] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 572.600207] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] result = hub.switch() [ 572.600207] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 572.600207] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] return self.greenlet.switch() [ 572.600207] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 572.600207] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] result = function(*args, **kwargs) [ 572.600207] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 572.600207] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] return func(*args, **kwargs) [ 572.600207] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 572.600207] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] raise e [ 572.600207] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 572.600207] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] nwinfo = self.network_api.allocate_for_instance( [ 572.600207] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 572.600207] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] created_port_ids = self._update_ports_for_instance( [ 572.600552] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 572.600552] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] with excutils.save_and_reraise_exception(): [ 572.600552] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.600552] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] self.force_reraise() [ 572.600552] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.600552] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] raise self.value [ 572.600552] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 572.600552] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] updated_port = self._update_port( [ 572.600552] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.600552] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] _ensure_no_port_binding_failure(port) [ 572.600552] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.600552] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] raise exception.PortBindingFailed(port_id=port['id']) [ 572.600954] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] nova.exception.PortBindingFailed: Binding failed for port 9add2eb2-422d-493e-ac3d-acf980ff6c90, please check neutron logs for more information. [ 572.600954] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] [ 572.600954] env[62965]: INFO nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Terminating instance [ 572.712764] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 2931c8b9-ed5f-4994-a4d5-50e1369d67e9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 572.912174] env[62965]: DEBUG nova.network.neutron [-] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.101430] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Acquiring lock "refresh_cache-2870d8cb-8c19-454e-8a6e-1675eba8ddd4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 573.101557] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Acquired lock "refresh_cache-2870d8cb-8c19-454e-8a6e-1675eba8ddd4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 573.101685] env[62965]: DEBUG nova.network.neutron [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 573.157627] env[62965]: DEBUG nova.compute.manager [req-ba0f2b60-3995-4e8a-ac90-e6f51820fe93 req-4bb5c817-a80a-4aac-b019-47a1f4f11977 service nova] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Received event network-changed-9add2eb2-422d-493e-ac3d-acf980ff6c90 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 573.157754] env[62965]: DEBUG nova.compute.manager [req-ba0f2b60-3995-4e8a-ac90-e6f51820fe93 req-4bb5c817-a80a-4aac-b019-47a1f4f11977 service nova] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Refreshing instance network info cache due to event network-changed-9add2eb2-422d-493e-ac3d-acf980ff6c90. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 573.157945] env[62965]: DEBUG oslo_concurrency.lockutils [req-ba0f2b60-3995-4e8a-ac90-e6f51820fe93 req-4bb5c817-a80a-4aac-b019-47a1f4f11977 service nova] Acquiring lock "refresh_cache-2870d8cb-8c19-454e-8a6e-1675eba8ddd4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 573.216456] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance f619dd08-0cdf-486f-8f13-707b254a8593 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 573.415414] env[62965]: INFO nova.compute.manager [-] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Took 1.22 seconds to deallocate network for instance. [ 573.626028] env[62965]: DEBUG nova.network.neutron [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 573.688240] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] Acquiring lock "f45f605f-cb91-4169-879c-bf9e6ce9c989" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.688240] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] Lock "f45f605f-cb91-4169-879c-bf9e6ce9c989" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.721111] env[62965]: DEBUG nova.network.neutron [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.721111] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 61702d54-59d2-4d13-9eed-59b052b7fe74 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 573.927558] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.222714] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Releasing lock "refresh_cache-2870d8cb-8c19-454e-8a6e-1675eba8ddd4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.222714] env[62965]: DEBUG nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 574.222714] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 574.223522] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 59187ca2-0482-4603-8689-542d2bf5d023 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 574.224671] env[62965]: DEBUG oslo_concurrency.lockutils [req-ba0f2b60-3995-4e8a-ac90-e6f51820fe93 req-4bb5c817-a80a-4aac-b019-47a1f4f11977 service nova] Acquired lock "refresh_cache-2870d8cb-8c19-454e-8a6e-1675eba8ddd4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 574.224853] env[62965]: DEBUG nova.network.neutron [req-ba0f2b60-3995-4e8a-ac90-e6f51820fe93 req-4bb5c817-a80a-4aac-b019-47a1f4f11977 service nova] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Refreshing network info cache for port 9add2eb2-422d-493e-ac3d-acf980ff6c90 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 574.225938] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-996c6af0-0548-4037-afed-4a3bfe1c6d6b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.237014] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cae22d03-a8cd-4e04-8504-b474d7b2a276 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.262162] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2870d8cb-8c19-454e-8a6e-1675eba8ddd4 could not be found. [ 574.262701] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 574.263414] env[62965]: INFO nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Took 0.04 seconds to destroy the instance on the hypervisor. [ 574.263414] env[62965]: DEBUG oslo.service.loopingcall [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 574.263545] env[62965]: DEBUG nova.compute.manager [-] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 574.263585] env[62965]: DEBUG nova.network.neutron [-] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 574.282736] env[62965]: DEBUG nova.network.neutron [-] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 574.409636] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Acquiring lock "02f113d2-f991-4268-b9a4-59e2e6bcf7a6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.409833] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Lock "02f113d2-f991-4268-b9a4-59e2e6bcf7a6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.732202] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 53519d3e-34d2-410f-ac19-69a015e407b9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 574.753835] env[62965]: DEBUG nova.network.neutron [req-ba0f2b60-3995-4e8a-ac90-e6f51820fe93 req-4bb5c817-a80a-4aac-b019-47a1f4f11977 service nova] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 574.786911] env[62965]: DEBUG nova.network.neutron [-] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.848385] env[62965]: DEBUG nova.network.neutron [req-ba0f2b60-3995-4e8a-ac90-e6f51820fe93 req-4bb5c817-a80a-4aac-b019-47a1f4f11977 service nova] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.237879] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance f594c50e-25b3-4e85-840c-7dfe0994a2c2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 575.289554] env[62965]: INFO nova.compute.manager [-] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Took 1.03 seconds to deallocate network for instance. [ 575.293033] env[62965]: DEBUG nova.compute.claims [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 575.293033] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.351286] env[62965]: DEBUG oslo_concurrency.lockutils [req-ba0f2b60-3995-4e8a-ac90-e6f51820fe93 req-4bb5c817-a80a-4aac-b019-47a1f4f11977 service nova] Releasing lock "refresh_cache-2870d8cb-8c19-454e-8a6e-1675eba8ddd4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 575.484842] env[62965]: DEBUG nova.compute.manager [req-9499581e-844a-4121-a3bf-9940268a588f req-00d759c0-0177-4289-9168-f14509680aca service nova] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Received event network-vif-deleted-9add2eb2-422d-493e-ac3d-acf980ff6c90 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 575.741483] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 09b62788-3082-4c99-9dfc-0843c4988feb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 576.244798] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 63c54ed4-4525-42f6-a99f-033f7b9838a7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 576.756058] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 222f547e-8fce-4576-bafc-809ea2ee8631 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 577.260633] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance af29af38-bd31-446a-a944-2bb43d9aa6c6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 577.766855] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 8a893e6e-37d4-4fd4-b0db-d6633117bdce has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 578.272179] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 8ac4264d-7d75-4c38-b8da-e3d90ffe9363 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 578.776275] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 971f72f8-2686-4ced-a2a1-b7ab8a1ea024 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 579.279545] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance eaf9ca94-4f45-403e-bb06-e9a9adf58f38 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 579.783455] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 79a4d3c8-ead6-4893-b0a7-89a8ff435e54 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 579.786443] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=62965) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 579.786657] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2048MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=62965) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 580.279637] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdd2af1d-0a07-43f9-bcc9-01e269b8a7f0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.295079] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3885ebdb-e1ba-48b1-9c70-bda95f4d9018 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.342973] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2daa1eee-7a10-466f-8254-1b0c4acc0c1a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.355119] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bdbe22e-10bd-4dbc-8d42-0375227a60ac {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.369504] env[62965]: DEBUG nova.compute.provider_tree [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 580.872659] env[62965]: DEBUG nova.scheduler.client.report [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 581.382029] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62965) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 581.382235] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 11.256s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 581.382479] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.362s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.456972] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd78adb0-bd26-476a-a45d-80196bbcc4c9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.472019] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66b0472d-b59b-4e8d-94da-c771de9bde90 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.512572] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dae362f-29bc-42bc-999d-3b1be9a1419c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.524455] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d260f151-efa1-4a71-b13f-2050d72eae2c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.542387] env[62965]: DEBUG nova.compute.provider_tree [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 583.045772] env[62965]: DEBUG nova.scheduler.client.report [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 583.141702] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Acquiring lock "4519d37a-f8e6-4a25-99a1-399925199a58" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.142037] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Lock "4519d37a-f8e6-4a25-99a1-399925199a58" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.554995] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.172s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 583.556363] env[62965]: ERROR nova.compute.manager [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fafa1864-43d6-483c-8a81-484a454f3f86, please check neutron logs for more information. [ 583.556363] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Traceback (most recent call last): [ 583.556363] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 583.556363] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] self.driver.spawn(context, instance, image_meta, [ 583.556363] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 583.556363] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 583.556363] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 583.556363] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] vm_ref = self.build_virtual_machine(instance, [ 583.556363] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 583.556363] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] vif_infos = vmwarevif.get_vif_info(self._session, [ 583.556363] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 583.556757] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] for vif in network_info: [ 583.556757] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 583.556757] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] return self._sync_wrapper(fn, *args, **kwargs) [ 583.556757] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 583.556757] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] self.wait() [ 583.556757] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 583.556757] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] self[:] = self._gt.wait() [ 583.556757] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 583.556757] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] return self._exit_event.wait() [ 583.556757] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 583.556757] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] result = hub.switch() [ 583.556757] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 583.556757] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] return self.greenlet.switch() [ 583.557157] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 583.557157] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] result = function(*args, **kwargs) [ 583.557157] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 583.557157] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] return func(*args, **kwargs) [ 583.557157] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 583.557157] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] raise e [ 583.557157] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 583.557157] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] nwinfo = self.network_api.allocate_for_instance( [ 583.557157] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 583.557157] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] created_port_ids = self._update_ports_for_instance( [ 583.557157] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 583.557157] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] with excutils.save_and_reraise_exception(): [ 583.557157] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 583.558754] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] self.force_reraise() [ 583.558754] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 583.558754] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] raise self.value [ 583.558754] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 583.558754] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] updated_port = self._update_port( [ 583.558754] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 583.558754] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] _ensure_no_port_binding_failure(port) [ 583.558754] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 583.558754] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] raise exception.PortBindingFailed(port_id=port['id']) [ 583.558754] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] nova.exception.PortBindingFailed: Binding failed for port fafa1864-43d6-483c-8a81-484a454f3f86, please check neutron logs for more information. [ 583.558754] env[62965]: ERROR nova.compute.manager [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] [ 583.559101] env[62965]: DEBUG nova.compute.utils [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Binding failed for port fafa1864-43d6-483c-8a81-484a454f3f86, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 583.559413] env[62965]: DEBUG oslo_concurrency.lockutils [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.254s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.565516] env[62965]: DEBUG nova.compute.manager [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Build of instance c46f89de-5156-49b4-b8f7-f60d78f1134b was re-scheduled: Binding failed for port fafa1864-43d6-483c-8a81-484a454f3f86, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 583.566165] env[62965]: DEBUG nova.compute.manager [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 583.566449] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Acquiring lock "refresh_cache-c46f89de-5156-49b4-b8f7-f60d78f1134b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 583.566606] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Acquired lock "refresh_cache-c46f89de-5156-49b4-b8f7-f60d78f1134b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 583.566769] env[62965]: DEBUG nova.network.neutron [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 584.142159] env[62965]: DEBUG nova.network.neutron [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 584.563670] env[62965]: DEBUG nova.network.neutron [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.658345] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dde9771-59c2-4098-83c0-47b2d269aaa7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.672083] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] Acquiring lock "6c7ed2a3-ed0f-476b-8232-d7a3c27adcee" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.672083] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] Lock "6c7ed2a3-ed0f-476b-8232-d7a3c27adcee" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.673018] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e382e099-c335-48a6-beb1-489ada0b893b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.708938] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-013411c1-b170-4ba0-8d34-1f0ea3b44e9c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.719539] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2988507-becf-496f-8138-f9b8c05288d5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.734286] env[62965]: DEBUG nova.compute.provider_tree [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 585.067219] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Releasing lock "refresh_cache-c46f89de-5156-49b4-b8f7-f60d78f1134b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 585.067520] env[62965]: DEBUG nova.compute.manager [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 585.067706] env[62965]: DEBUG nova.compute.manager [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 585.067917] env[62965]: DEBUG nova.network.neutron [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 585.110121] env[62965]: DEBUG nova.network.neutron [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 585.240595] env[62965]: DEBUG nova.scheduler.client.report [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 585.614537] env[62965]: DEBUG nova.network.neutron [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.747852] env[62965]: DEBUG oslo_concurrency.lockutils [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.188s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 585.748561] env[62965]: ERROR nova.compute.manager [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7ca66b62-dbd5-44c0-8bef-3ca7cf104ca1, please check neutron logs for more information. [ 585.748561] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Traceback (most recent call last): [ 585.748561] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 585.748561] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] self.driver.spawn(context, instance, image_meta, [ 585.748561] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 585.748561] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 585.748561] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 585.748561] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] vm_ref = self.build_virtual_machine(instance, [ 585.748561] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 585.748561] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] vif_infos = vmwarevif.get_vif_info(self._session, [ 585.748561] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 585.748942] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] for vif in network_info: [ 585.748942] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 585.748942] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] return self._sync_wrapper(fn, *args, **kwargs) [ 585.748942] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 585.748942] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] self.wait() [ 585.748942] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 585.748942] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] self[:] = self._gt.wait() [ 585.748942] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 585.748942] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] return self._exit_event.wait() [ 585.748942] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 585.748942] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] result = hub.switch() [ 585.748942] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 585.748942] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] return self.greenlet.switch() [ 585.750055] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 585.750055] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] result = function(*args, **kwargs) [ 585.750055] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 585.750055] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] return func(*args, **kwargs) [ 585.750055] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 585.750055] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] raise e [ 585.750055] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 585.750055] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] nwinfo = self.network_api.allocate_for_instance( [ 585.750055] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 585.750055] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] created_port_ids = self._update_ports_for_instance( [ 585.750055] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 585.750055] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] with excutils.save_and_reraise_exception(): [ 585.750055] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.750484] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] self.force_reraise() [ 585.750484] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.750484] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] raise self.value [ 585.750484] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 585.750484] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] updated_port = self._update_port( [ 585.750484] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.750484] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] _ensure_no_port_binding_failure(port) [ 585.750484] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.750484] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] raise exception.PortBindingFailed(port_id=port['id']) [ 585.750484] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] nova.exception.PortBindingFailed: Binding failed for port 7ca66b62-dbd5-44c0-8bef-3ca7cf104ca1, please check neutron logs for more information. [ 585.750484] env[62965]: ERROR nova.compute.manager [instance: 709bac33-9605-4254-bf38-49fc7892b15e] [ 585.750834] env[62965]: DEBUG nova.compute.utils [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Binding failed for port 7ca66b62-dbd5-44c0-8bef-3ca7cf104ca1, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 585.750834] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.557s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.754718] env[62965]: DEBUG nova.compute.manager [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Build of instance 709bac33-9605-4254-bf38-49fc7892b15e was re-scheduled: Binding failed for port 7ca66b62-dbd5-44c0-8bef-3ca7cf104ca1, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 585.754904] env[62965]: DEBUG nova.compute.manager [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 585.755144] env[62965]: DEBUG oslo_concurrency.lockutils [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] Acquiring lock "refresh_cache-709bac33-9605-4254-bf38-49fc7892b15e" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 585.755291] env[62965]: DEBUG oslo_concurrency.lockutils [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] Acquired lock "refresh_cache-709bac33-9605-4254-bf38-49fc7892b15e" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.755449] env[62965]: DEBUG nova.network.neutron [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 586.120616] env[62965]: INFO nova.compute.manager [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: c46f89de-5156-49b4-b8f7-f60d78f1134b] Took 1.05 seconds to deallocate network for instance. [ 586.282022] env[62965]: DEBUG nova.network.neutron [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 586.358037] env[62965]: DEBUG nova.network.neutron [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.401047] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] Acquiring lock "852e7cdf-5da3-4217-8960-f3204d4c1035" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 586.401294] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] Lock "852e7cdf-5da3-4217-8960-f3204d4c1035" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.779773] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44342dc7-0281-40a4-8a45-e3d76a3601d9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.790057] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab9630a4-2f95-4708-96c8-975109608a5b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.834313] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86641add-26ad-46ff-b140-2b2084eab53c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.841981] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f07e9b8-6abf-48fe-b292-6ce5e9612e50 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.856279] env[62965]: DEBUG nova.compute.provider_tree [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 586.861406] env[62965]: DEBUG oslo_concurrency.lockutils [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] Acquiring lock "b2f38e92-37a1-4e0b-9292-b20ea43a2038" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 586.861624] env[62965]: DEBUG oslo_concurrency.lockutils [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] Lock "b2f38e92-37a1-4e0b-9292-b20ea43a2038" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.861968] env[62965]: DEBUG oslo_concurrency.lockutils [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] Releasing lock "refresh_cache-709bac33-9605-4254-bf38-49fc7892b15e" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 586.862167] env[62965]: DEBUG nova.compute.manager [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 586.862323] env[62965]: DEBUG nova.compute.manager [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 586.862480] env[62965]: DEBUG nova.network.neutron [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 586.879162] env[62965]: DEBUG nova.network.neutron [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 587.177493] env[62965]: INFO nova.scheduler.client.report [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Deleted allocations for instance c46f89de-5156-49b4-b8f7-f60d78f1134b [ 587.354535] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "83c4b731-780a-4f3c-bbdd-b624a52365ef" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.355012] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "83c4b731-780a-4f3c-bbdd-b624a52365ef" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.361861] env[62965]: DEBUG nova.scheduler.client.report [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 587.385148] env[62965]: DEBUG nova.network.neutron [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.690180] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f800f776-8869-4275-8704-143f2d9c4a2a tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Lock "c46f89de-5156-49b4-b8f7-f60d78f1134b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.411s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 587.770473] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] Acquiring lock "f8ba32e4-6234-4dee-8bc5-b7a12105a3c4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.771194] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] Lock "f8ba32e4-6234-4dee-8bc5-b7a12105a3c4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.868320] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.117s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 587.869139] env[62965]: ERROR nova.compute.manager [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 688eba8b-b5ba-4d6b-9baf-2300a4d2dac2, please check neutron logs for more information. [ 587.869139] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Traceback (most recent call last): [ 587.869139] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 587.869139] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] self.driver.spawn(context, instance, image_meta, [ 587.869139] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 587.869139] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] self._vmops.spawn(context, instance, image_meta, injected_files, [ 587.869139] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 587.869139] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] vm_ref = self.build_virtual_machine(instance, [ 587.869139] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 587.869139] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] vif_infos = vmwarevif.get_vif_info(self._session, [ 587.869139] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 587.869490] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] for vif in network_info: [ 587.869490] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 587.869490] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] return self._sync_wrapper(fn, *args, **kwargs) [ 587.869490] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 587.869490] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] self.wait() [ 587.869490] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 587.869490] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] self[:] = self._gt.wait() [ 587.869490] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 587.869490] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] return self._exit_event.wait() [ 587.869490] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 587.869490] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] result = hub.switch() [ 587.869490] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 587.869490] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] return self.greenlet.switch() [ 587.869913] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 587.869913] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] result = function(*args, **kwargs) [ 587.869913] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 587.869913] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] return func(*args, **kwargs) [ 587.869913] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 587.869913] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] raise e [ 587.869913] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 587.869913] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] nwinfo = self.network_api.allocate_for_instance( [ 587.869913] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 587.869913] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] created_port_ids = self._update_ports_for_instance( [ 587.869913] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 587.869913] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] with excutils.save_and_reraise_exception(): [ 587.869913] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.870358] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] self.force_reraise() [ 587.870358] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.870358] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] raise self.value [ 587.870358] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 587.870358] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] updated_port = self._update_port( [ 587.870358] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.870358] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] _ensure_no_port_binding_failure(port) [ 587.870358] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.870358] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] raise exception.PortBindingFailed(port_id=port['id']) [ 587.870358] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] nova.exception.PortBindingFailed: Binding failed for port 688eba8b-b5ba-4d6b-9baf-2300a4d2dac2, please check neutron logs for more information. [ 587.870358] env[62965]: ERROR nova.compute.manager [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] [ 587.870750] env[62965]: DEBUG nova.compute.utils [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Binding failed for port 688eba8b-b5ba-4d6b-9baf-2300a4d2dac2, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 587.872027] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.152s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.874710] env[62965]: DEBUG nova.compute.manager [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Build of instance 7780b346-28a4-4d80-9f83-a452c23741ec was re-scheduled: Binding failed for port 688eba8b-b5ba-4d6b-9baf-2300a4d2dac2, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 587.875728] env[62965]: DEBUG nova.compute.manager [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 587.875728] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] Acquiring lock "refresh_cache-7780b346-28a4-4d80-9f83-a452c23741ec" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 587.875728] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] Acquired lock "refresh_cache-7780b346-28a4-4d80-9f83-a452c23741ec" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 587.875728] env[62965]: DEBUG nova.network.neutron [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 587.886761] env[62965]: INFO nova.compute.manager [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] [instance: 709bac33-9605-4254-bf38-49fc7892b15e] Took 1.02 seconds to deallocate network for instance. [ 588.195293] env[62965]: DEBUG nova.compute.manager [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 588.326514] env[62965]: DEBUG oslo_concurrency.lockutils [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Acquiring lock "ece118fc-a39e-443a-bfd1-fc332da2452f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.326746] env[62965]: DEBUG oslo_concurrency.lockutils [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Lock "ece118fc-a39e-443a-bfd1-fc332da2452f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.400737] env[62965]: DEBUG nova.network.neutron [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 588.512035] env[62965]: DEBUG nova.network.neutron [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.717859] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.816023] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e882a86d-a429-4b1b-bbba-b9441ccaf554 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.823537] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07bda581-2c1e-4fe5-a61a-af5c418f80ee {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.856974] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35df009e-2ff6-41a1-9a93-88088a45f1ab {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.864619] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04d9aec0-fd50-4cb5-93ae-5cd0a4d171f2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.878071] env[62965]: DEBUG nova.compute.provider_tree [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 588.920523] env[62965]: INFO nova.scheduler.client.report [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] Deleted allocations for instance 709bac33-9605-4254-bf38-49fc7892b15e [ 589.015518] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] Releasing lock "refresh_cache-7780b346-28a4-4d80-9f83-a452c23741ec" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.015518] env[62965]: DEBUG nova.compute.manager [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 589.015518] env[62965]: DEBUG nova.compute.manager [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 589.015518] env[62965]: DEBUG nova.network.neutron [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 589.032487] env[62965]: DEBUG nova.network.neutron [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 589.381844] env[62965]: DEBUG nova.scheduler.client.report [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 589.430588] env[62965]: DEBUG oslo_concurrency.lockutils [None req-732e6720-2b8b-4c6a-89b6-9d5211d14089 tempest-ServerDiagnosticsTest-1710154591 tempest-ServerDiagnosticsTest-1710154591-project-member] Lock "709bac33-9605-4254-bf38-49fc7892b15e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.943s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 589.535878] env[62965]: DEBUG nova.network.neutron [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.890949] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.019s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 589.891620] env[62965]: ERROR nova.compute.manager [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d43a2c0d-b31a-4ee0-bccb-821a706f2c80, please check neutron logs for more information. [ 589.891620] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Traceback (most recent call last): [ 589.891620] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 589.891620] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] self.driver.spawn(context, instance, image_meta, [ 589.891620] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 589.891620] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] self._vmops.spawn(context, instance, image_meta, injected_files, [ 589.891620] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 589.891620] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] vm_ref = self.build_virtual_machine(instance, [ 589.891620] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 589.891620] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] vif_infos = vmwarevif.get_vif_info(self._session, [ 589.891620] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 589.892211] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] for vif in network_info: [ 589.892211] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 589.892211] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] return self._sync_wrapper(fn, *args, **kwargs) [ 589.892211] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 589.892211] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] self.wait() [ 589.892211] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 589.892211] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] self[:] = self._gt.wait() [ 589.892211] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 589.892211] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] return self._exit_event.wait() [ 589.892211] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 589.892211] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] result = hub.switch() [ 589.892211] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 589.892211] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] return self.greenlet.switch() [ 589.892572] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 589.892572] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] result = function(*args, **kwargs) [ 589.892572] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 589.892572] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] return func(*args, **kwargs) [ 589.892572] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 589.892572] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] raise e [ 589.892572] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 589.892572] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] nwinfo = self.network_api.allocate_for_instance( [ 589.892572] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 589.892572] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] created_port_ids = self._update_ports_for_instance( [ 589.892572] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 589.892572] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] with excutils.save_and_reraise_exception(): [ 589.892572] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.892955] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] self.force_reraise() [ 589.892955] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.892955] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] raise self.value [ 589.892955] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 589.892955] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] updated_port = self._update_port( [ 589.892955] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.892955] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] _ensure_no_port_binding_failure(port) [ 589.892955] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.892955] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] raise exception.PortBindingFailed(port_id=port['id']) [ 589.892955] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] nova.exception.PortBindingFailed: Binding failed for port d43a2c0d-b31a-4ee0-bccb-821a706f2c80, please check neutron logs for more information. [ 589.892955] env[62965]: ERROR nova.compute.manager [instance: fa55e796-81d0-46b7-b025-9119bae2f305] [ 589.893292] env[62965]: DEBUG nova.compute.utils [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Binding failed for port d43a2c0d-b31a-4ee0-bccb-821a706f2c80, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 589.893538] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.864s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.899908] env[62965]: DEBUG nova.compute.manager [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Build of instance fa55e796-81d0-46b7-b025-9119bae2f305 was re-scheduled: Binding failed for port d43a2c0d-b31a-4ee0-bccb-821a706f2c80, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 589.900414] env[62965]: DEBUG nova.compute.manager [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 589.900687] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Acquiring lock "refresh_cache-fa55e796-81d0-46b7-b025-9119bae2f305" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 589.900881] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Acquired lock "refresh_cache-fa55e796-81d0-46b7-b025-9119bae2f305" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 589.904019] env[62965]: DEBUG nova.network.neutron [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 589.933549] env[62965]: DEBUG nova.compute.manager [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 590.037642] env[62965]: INFO nova.compute.manager [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] [instance: 7780b346-28a4-4d80-9f83-a452c23741ec] Took 1.02 seconds to deallocate network for instance. [ 590.427028] env[62965]: DEBUG nova.network.neutron [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 590.460577] env[62965]: DEBUG oslo_concurrency.lockutils [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.581750] env[62965]: DEBUG nova.network.neutron [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.924207] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b0a9c8f-9b1a-4cb2-aa3f-90db0cbc26fc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.939020] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ddfe47f-373f-4359-ba9b-36ccdb3278f0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.969099] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e275a52a-e14e-41bb-8be0-2abbd92f0b01 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.978142] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c04f908b-6518-44ee-bb05-7786638c5cc6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.997551] env[62965]: DEBUG nova.compute.provider_tree [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 591.078078] env[62965]: INFO nova.scheduler.client.report [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] Deleted allocations for instance 7780b346-28a4-4d80-9f83-a452c23741ec [ 591.084486] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Releasing lock "refresh_cache-fa55e796-81d0-46b7-b025-9119bae2f305" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 591.084772] env[62965]: DEBUG nova.compute.manager [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 591.085070] env[62965]: DEBUG nova.compute.manager [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 591.085268] env[62965]: DEBUG nova.network.neutron [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 591.119263] env[62965]: DEBUG nova.network.neutron [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 591.500677] env[62965]: DEBUG nova.scheduler.client.report [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 591.588866] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1c8cd555-495e-47da-a51e-4dd2dd54f2aa tempest-ServerActionsTestOtherA-201298064 tempest-ServerActionsTestOtherA-201298064-project-member] Lock "7780b346-28a4-4d80-9f83-a452c23741ec" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.762s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 591.620998] env[62965]: DEBUG nova.network.neutron [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.007795] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.114s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 592.008763] env[62965]: ERROR nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d49a2434-0352-4b6f-95a8-3dbf7c5ef1cd, please check neutron logs for more information. [ 592.008763] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Traceback (most recent call last): [ 592.008763] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 592.008763] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] self.driver.spawn(context, instance, image_meta, [ 592.008763] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 592.008763] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 592.008763] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 592.008763] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] vm_ref = self.build_virtual_machine(instance, [ 592.008763] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 592.008763] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] vif_infos = vmwarevif.get_vif_info(self._session, [ 592.008763] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 592.009194] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] for vif in network_info: [ 592.009194] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 592.009194] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] return self._sync_wrapper(fn, *args, **kwargs) [ 592.009194] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 592.009194] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] self.wait() [ 592.009194] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 592.009194] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] self[:] = self._gt.wait() [ 592.009194] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 592.009194] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] return self._exit_event.wait() [ 592.009194] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 592.009194] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] result = hub.switch() [ 592.009194] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 592.009194] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] return self.greenlet.switch() [ 592.009567] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 592.009567] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] result = function(*args, **kwargs) [ 592.009567] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 592.009567] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] return func(*args, **kwargs) [ 592.009567] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 592.009567] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] raise e [ 592.009567] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 592.009567] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] nwinfo = self.network_api.allocate_for_instance( [ 592.009567] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 592.009567] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] created_port_ids = self._update_ports_for_instance( [ 592.009567] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 592.009567] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] with excutils.save_and_reraise_exception(): [ 592.009567] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.009946] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] self.force_reraise() [ 592.009946] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.009946] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] raise self.value [ 592.009946] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 592.009946] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] updated_port = self._update_port( [ 592.009946] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.009946] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] _ensure_no_port_binding_failure(port) [ 592.009946] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.009946] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] raise exception.PortBindingFailed(port_id=port['id']) [ 592.009946] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] nova.exception.PortBindingFailed: Binding failed for port d49a2434-0352-4b6f-95a8-3dbf7c5ef1cd, please check neutron logs for more information. [ 592.009946] env[62965]: ERROR nova.compute.manager [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] [ 592.010278] env[62965]: DEBUG nova.compute.utils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Binding failed for port d49a2434-0352-4b6f-95a8-3dbf7c5ef1cd, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 592.011034] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.718s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.012836] env[62965]: INFO nova.compute.claims [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 592.017859] env[62965]: DEBUG nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Build of instance b811920a-14a2-4c11-8b13-78d920f08a0e was re-scheduled: Binding failed for port d49a2434-0352-4b6f-95a8-3dbf7c5ef1cd, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 592.017859] env[62965]: DEBUG nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 592.017859] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Acquiring lock "refresh_cache-b811920a-14a2-4c11-8b13-78d920f08a0e" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 592.017859] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Acquired lock "refresh_cache-b811920a-14a2-4c11-8b13-78d920f08a0e" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 592.018195] env[62965]: DEBUG nova.network.neutron [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 592.093890] env[62965]: DEBUG nova.compute.manager [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 592.123967] env[62965]: INFO nova.compute.manager [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] [instance: fa55e796-81d0-46b7-b025-9119bae2f305] Took 1.04 seconds to deallocate network for instance. [ 592.548264] env[62965]: DEBUG nova.network.neutron [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 592.620346] env[62965]: DEBUG oslo_concurrency.lockutils [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.666497] env[62965]: DEBUG nova.network.neutron [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.170770] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Releasing lock "refresh_cache-b811920a-14a2-4c11-8b13-78d920f08a0e" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 593.171076] env[62965]: DEBUG nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 593.171314] env[62965]: DEBUG nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 593.171542] env[62965]: DEBUG nova.network.neutron [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 593.175210] env[62965]: INFO nova.scheduler.client.report [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Deleted allocations for instance fa55e796-81d0-46b7-b025-9119bae2f305 [ 593.239815] env[62965]: DEBUG nova.network.neutron [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 593.581350] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75bdc094-f6f8-4664-8498-0710f3cc2a44 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.590567] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c20163a9-be69-4718-b634-ed8f4fe15dca {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.626401] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64446402-b3ff-434a-8a77-9e2f717d0ed5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.632446] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf8b4115-f3d8-4dc6-b99a-e4d847f32f6f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.648765] env[62965]: DEBUG nova.compute.provider_tree [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 593.691755] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b51065bd-e8ae-4bd0-aea4-f9354a3fd40e tempest-ServersAdminTestJSON-715792014 tempest-ServersAdminTestJSON-715792014-project-member] Lock "fa55e796-81d0-46b7-b025-9119bae2f305" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.494s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 593.744050] env[62965]: DEBUG nova.network.neutron [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.156078] env[62965]: DEBUG nova.scheduler.client.report [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 594.195347] env[62965]: DEBUG nova.compute.manager [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 594.217094] env[62965]: DEBUG oslo_concurrency.lockutils [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] Acquiring lock "1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.217094] env[62965]: DEBUG oslo_concurrency.lockutils [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] Lock "1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.250217] env[62965]: INFO nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: b811920a-14a2-4c11-8b13-78d920f08a0e] Took 1.08 seconds to deallocate network for instance. [ 594.661798] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.651s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 594.662418] env[62965]: DEBUG nova.compute.manager [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 594.667083] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.339s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.731669] env[62965]: DEBUG oslo_concurrency.lockutils [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.177189] env[62965]: DEBUG nova.compute.utils [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 595.181655] env[62965]: DEBUG nova.compute.manager [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 595.182027] env[62965]: DEBUG nova.network.neutron [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 595.247576] env[62965]: DEBUG nova.policy [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f50a3f6cb8574392890da5144157d996', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fb1907c8137f4bc1ae2b8337d3606b53', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 595.286532] env[62965]: INFO nova.scheduler.client.report [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Deleted allocations for instance b811920a-14a2-4c11-8b13-78d920f08a0e [ 595.642037] env[62965]: DEBUG nova.network.neutron [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Successfully created port: 3211a0c8-0637-4db2-befc-3acd54e53435 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 595.687321] env[62965]: DEBUG nova.compute.manager [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 595.754259] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c030d3e-ecc6-4fab-8132-97f1dd03d24f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.768068] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f27d1295-76c7-4357-bdca-ee6943f6724d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.802659] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Lock "b811920a-14a2-4c11-8b13-78d920f08a0e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.155s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.805180] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae886e51-555f-4e92-b7ba-9df31919e441 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.813406] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11768fd9-1fce-4493-a1be-be42f86fe54f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.832707] env[62965]: DEBUG nova.compute.provider_tree [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 596.310023] env[62965]: DEBUG nova.compute.manager [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 596.335079] env[62965]: DEBUG nova.scheduler.client.report [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 596.701671] env[62965]: DEBUG nova.compute.manager [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 596.738226] env[62965]: DEBUG nova.virt.hardware [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 596.738226] env[62965]: DEBUG nova.virt.hardware [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 596.738226] env[62965]: DEBUG nova.virt.hardware [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 596.738394] env[62965]: DEBUG nova.virt.hardware [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 596.738394] env[62965]: DEBUG nova.virt.hardware [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 596.738394] env[62965]: DEBUG nova.virt.hardware [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 596.738558] env[62965]: DEBUG nova.virt.hardware [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 596.738638] env[62965]: DEBUG nova.virt.hardware [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 596.738784] env[62965]: DEBUG nova.virt.hardware [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 596.739730] env[62965]: DEBUG nova.virt.hardware [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 596.739958] env[62965]: DEBUG nova.virt.hardware [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 596.740861] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5b4f364-1136-48e2-9a2b-28e58f45cc14 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.749361] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0256f6e7-9c55-425d-8ae0-8c3db84b2a34 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.838724] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.842962] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.176s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 596.843567] env[62965]: ERROR nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3c249e6a-17cb-4734-b532-233e86c66d5c, please check neutron logs for more information. [ 596.843567] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Traceback (most recent call last): [ 596.843567] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 596.843567] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] self.driver.spawn(context, instance, image_meta, [ 596.843567] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 596.843567] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 596.843567] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 596.843567] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] vm_ref = self.build_virtual_machine(instance, [ 596.843567] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 596.843567] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] vif_infos = vmwarevif.get_vif_info(self._session, [ 596.843567] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 596.843934] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] for vif in network_info: [ 596.843934] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 596.843934] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] return self._sync_wrapper(fn, *args, **kwargs) [ 596.843934] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 596.843934] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] self.wait() [ 596.843934] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 596.843934] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] self[:] = self._gt.wait() [ 596.843934] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 596.843934] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] return self._exit_event.wait() [ 596.843934] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 596.843934] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] result = hub.switch() [ 596.843934] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 596.843934] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] return self.greenlet.switch() [ 596.844327] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 596.844327] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] result = function(*args, **kwargs) [ 596.844327] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 596.844327] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] return func(*args, **kwargs) [ 596.844327] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 596.844327] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] raise e [ 596.844327] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 596.844327] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] nwinfo = self.network_api.allocate_for_instance( [ 596.844327] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 596.844327] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] created_port_ids = self._update_ports_for_instance( [ 596.844327] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 596.844327] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] with excutils.save_and_reraise_exception(): [ 596.844327] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.844722] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] self.force_reraise() [ 596.844722] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.844722] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] raise self.value [ 596.844722] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 596.844722] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] updated_port = self._update_port( [ 596.844722] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.844722] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] _ensure_no_port_binding_failure(port) [ 596.844722] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.844722] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] raise exception.PortBindingFailed(port_id=port['id']) [ 596.844722] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] nova.exception.PortBindingFailed: Binding failed for port 3c249e6a-17cb-4734-b532-233e86c66d5c, please check neutron logs for more information. [ 596.844722] env[62965]: ERROR nova.compute.manager [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] [ 596.845173] env[62965]: DEBUG nova.compute.utils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Binding failed for port 3c249e6a-17cb-4734-b532-233e86c66d5c, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 596.845949] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.501s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.847824] env[62965]: INFO nova.compute.claims [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 596.850999] env[62965]: DEBUG nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Build of instance 4a75ddc4-add2-45dc-b086-bc163990db4f was re-scheduled: Binding failed for port 3c249e6a-17cb-4734-b532-233e86c66d5c, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 596.851459] env[62965]: DEBUG nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 596.851724] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Acquiring lock "refresh_cache-4a75ddc4-add2-45dc-b086-bc163990db4f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 596.851996] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Acquired lock "refresh_cache-4a75ddc4-add2-45dc-b086-bc163990db4f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 596.852143] env[62965]: DEBUG nova.network.neutron [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 597.072163] env[62965]: ERROR nova.compute.manager [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3211a0c8-0637-4db2-befc-3acd54e53435, please check neutron logs for more information. [ 597.072163] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 597.072163] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 597.072163] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 597.072163] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 597.072163] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 597.072163] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 597.072163] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 597.072163] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.072163] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 597.072163] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.072163] env[62965]: ERROR nova.compute.manager raise self.value [ 597.072163] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 597.072163] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 597.072163] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.072163] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 597.072868] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.072868] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 597.072868] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3211a0c8-0637-4db2-befc-3acd54e53435, please check neutron logs for more information. [ 597.072868] env[62965]: ERROR nova.compute.manager [ 597.072868] env[62965]: Traceback (most recent call last): [ 597.072868] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 597.072868] env[62965]: listener.cb(fileno) [ 597.072868] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.072868] env[62965]: result = function(*args, **kwargs) [ 597.072868] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 597.072868] env[62965]: return func(*args, **kwargs) [ 597.072868] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 597.072868] env[62965]: raise e [ 597.072868] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 597.072868] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 597.072868] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 597.072868] env[62965]: created_port_ids = self._update_ports_for_instance( [ 597.072868] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 597.072868] env[62965]: with excutils.save_and_reraise_exception(): [ 597.072868] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.072868] env[62965]: self.force_reraise() [ 597.072868] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.072868] env[62965]: raise self.value [ 597.072868] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 597.072868] env[62965]: updated_port = self._update_port( [ 597.072868] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.072868] env[62965]: _ensure_no_port_binding_failure(port) [ 597.072868] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.072868] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 597.074075] env[62965]: nova.exception.PortBindingFailed: Binding failed for port 3211a0c8-0637-4db2-befc-3acd54e53435, please check neutron logs for more information. [ 597.074075] env[62965]: Removing descriptor: 19 [ 597.074075] env[62965]: ERROR nova.compute.manager [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3211a0c8-0637-4db2-befc-3acd54e53435, please check neutron logs for more information. [ 597.074075] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Traceback (most recent call last): [ 597.074075] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 597.074075] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] yield resources [ 597.074075] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 597.074075] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] self.driver.spawn(context, instance, image_meta, [ 597.074075] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 597.074075] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 597.074075] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 597.074075] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] vm_ref = self.build_virtual_machine(instance, [ 597.074655] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 597.074655] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] vif_infos = vmwarevif.get_vif_info(self._session, [ 597.074655] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 597.074655] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] for vif in network_info: [ 597.074655] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 597.074655] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] return self._sync_wrapper(fn, *args, **kwargs) [ 597.074655] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 597.074655] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] self.wait() [ 597.074655] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 597.074655] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] self[:] = self._gt.wait() [ 597.074655] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 597.074655] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] return self._exit_event.wait() [ 597.074655] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 597.076188] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] result = hub.switch() [ 597.076188] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 597.076188] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] return self.greenlet.switch() [ 597.076188] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.076188] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] result = function(*args, **kwargs) [ 597.076188] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 597.076188] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] return func(*args, **kwargs) [ 597.076188] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 597.076188] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] raise e [ 597.076188] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 597.076188] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] nwinfo = self.network_api.allocate_for_instance( [ 597.076188] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 597.076188] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] created_port_ids = self._update_ports_for_instance( [ 597.076635] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 597.076635] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] with excutils.save_and_reraise_exception(): [ 597.076635] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.076635] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] self.force_reraise() [ 597.076635] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.076635] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] raise self.value [ 597.076635] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 597.076635] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] updated_port = self._update_port( [ 597.076635] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.076635] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] _ensure_no_port_binding_failure(port) [ 597.076635] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.076635] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] raise exception.PortBindingFailed(port_id=port['id']) [ 597.077246] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] nova.exception.PortBindingFailed: Binding failed for port 3211a0c8-0637-4db2-befc-3acd54e53435, please check neutron logs for more information. [ 597.077246] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] [ 597.077246] env[62965]: INFO nova.compute.manager [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Terminating instance [ 597.378149] env[62965]: DEBUG nova.network.neutron [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 597.448938] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] Acquiring lock "99db6bef-6501-4ab3-a329-0877ca303f1b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.450339] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] Lock "99db6bef-6501-4ab3-a329-0877ca303f1b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.472870] env[62965]: DEBUG nova.network.neutron [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.581864] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Acquiring lock "refresh_cache-647a3ba2-8b84-45f9-9d28-130eb0521d1b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.582112] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Acquired lock "refresh_cache-647a3ba2-8b84-45f9-9d28-130eb0521d1b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.582286] env[62965]: DEBUG nova.network.neutron [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 597.769191] env[62965]: DEBUG nova.compute.manager [req-3669f799-c2ae-4359-badb-38ae3701e8b0 req-37ce53ac-f30c-41c0-a03a-fe501caa0e81 service nova] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Received event network-changed-3211a0c8-0637-4db2-befc-3acd54e53435 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 597.769412] env[62965]: DEBUG nova.compute.manager [req-3669f799-c2ae-4359-badb-38ae3701e8b0 req-37ce53ac-f30c-41c0-a03a-fe501caa0e81 service nova] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Refreshing instance network info cache due to event network-changed-3211a0c8-0637-4db2-befc-3acd54e53435. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 597.769682] env[62965]: DEBUG oslo_concurrency.lockutils [req-3669f799-c2ae-4359-badb-38ae3701e8b0 req-37ce53ac-f30c-41c0-a03a-fe501caa0e81 service nova] Acquiring lock "refresh_cache-647a3ba2-8b84-45f9-9d28-130eb0521d1b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.976090] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Releasing lock "refresh_cache-4a75ddc4-add2-45dc-b086-bc163990db4f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 597.976582] env[62965]: DEBUG nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 597.976850] env[62965]: DEBUG nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 597.978413] env[62965]: DEBUG nova.network.neutron [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 598.000786] env[62965]: DEBUG nova.network.neutron [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 598.192544] env[62965]: DEBUG nova.network.neutron [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 598.347790] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed7d0853-c06d-44d4-9955-fb3d0adeab33 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.355770] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7248f22f-8645-40fc-8b2d-40c67fdfe676 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.388356] env[62965]: DEBUG nova.network.neutron [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.390538] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb5b50ee-9407-48d8-9c09-186936281bda {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.400550] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e0ec738-dced-4b1d-8f79-aa404a17b505 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.415643] env[62965]: DEBUG nova.compute.provider_tree [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 598.505266] env[62965]: DEBUG nova.network.neutron [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.897146] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Releasing lock "refresh_cache-647a3ba2-8b84-45f9-9d28-130eb0521d1b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.897146] env[62965]: DEBUG nova.compute.manager [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 598.897146] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 598.897146] env[62965]: DEBUG oslo_concurrency.lockutils [req-3669f799-c2ae-4359-badb-38ae3701e8b0 req-37ce53ac-f30c-41c0-a03a-fe501caa0e81 service nova] Acquired lock "refresh_cache-647a3ba2-8b84-45f9-9d28-130eb0521d1b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.897146] env[62965]: DEBUG nova.network.neutron [req-3669f799-c2ae-4359-badb-38ae3701e8b0 req-37ce53ac-f30c-41c0-a03a-fe501caa0e81 service nova] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Refreshing network info cache for port 3211a0c8-0637-4db2-befc-3acd54e53435 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 598.898377] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6352dd07-2415-4fc7-875e-386df69a6e05 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.909220] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc23fc90-668e-4074-8f7c-ba10ecee2ef1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.922695] env[62965]: DEBUG nova.scheduler.client.report [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 598.938460] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 647a3ba2-8b84-45f9-9d28-130eb0521d1b could not be found. [ 598.938856] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 598.939110] env[62965]: INFO nova.compute.manager [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 598.939404] env[62965]: DEBUG oslo.service.loopingcall [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 598.940073] env[62965]: DEBUG nova.compute.manager [-] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 598.941035] env[62965]: DEBUG nova.network.neutron [-] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 598.959360] env[62965]: DEBUG nova.network.neutron [-] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 599.008809] env[62965]: INFO nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 4a75ddc4-add2-45dc-b086-bc163990db4f] Took 1.03 seconds to deallocate network for instance. [ 599.419910] env[62965]: DEBUG nova.network.neutron [req-3669f799-c2ae-4359-badb-38ae3701e8b0 req-37ce53ac-f30c-41c0-a03a-fe501caa0e81 service nova] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 599.428689] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.583s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 599.429312] env[62965]: DEBUG nova.compute.manager [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 599.432058] env[62965]: DEBUG oslo_concurrency.lockutils [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.053s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.433540] env[62965]: INFO nova.compute.claims [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 599.461239] env[62965]: DEBUG nova.network.neutron [-] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.518020] env[62965]: DEBUG nova.network.neutron [req-3669f799-c2ae-4359-badb-38ae3701e8b0 req-37ce53ac-f30c-41c0-a03a-fe501caa0e81 service nova] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.933184] env[62965]: DEBUG nova.compute.manager [req-1f3b22ae-f020-41be-9d2a-abd13c77766a req-11b213f0-119f-4f93-86d6-17c35deaf1b1 service nova] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Received event network-vif-deleted-3211a0c8-0637-4db2-befc-3acd54e53435 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 599.941017] env[62965]: DEBUG nova.compute.utils [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 599.945545] env[62965]: DEBUG nova.compute.manager [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 599.945897] env[62965]: DEBUG nova.network.neutron [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 599.967791] env[62965]: INFO nova.compute.manager [-] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Took 1.03 seconds to deallocate network for instance. [ 599.968574] env[62965]: DEBUG nova.compute.claims [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 599.968574] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.005488] env[62965]: DEBUG nova.policy [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6614b29c1a5b4571b3d423a371f004ec', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9e00ea3dce504db6b1eade648a13ea2d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 600.025940] env[62965]: DEBUG oslo_concurrency.lockutils [req-3669f799-c2ae-4359-badb-38ae3701e8b0 req-37ce53ac-f30c-41c0-a03a-fe501caa0e81 service nova] Releasing lock "refresh_cache-647a3ba2-8b84-45f9-9d28-130eb0521d1b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 600.055806] env[62965]: INFO nova.scheduler.client.report [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Deleted allocations for instance 4a75ddc4-add2-45dc-b086-bc163990db4f [ 600.448939] env[62965]: DEBUG nova.compute.manager [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 600.471764] env[62965]: DEBUG nova.network.neutron [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Successfully created port: 3974d385-edd8-4d22-a7c2-1f79a1c724ec {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 600.564559] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Lock "4a75ddc4-add2-45dc-b086-bc163990db4f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.848s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.964265] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f175eee-04ef-42d6-b945-aac36e677bd3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.979875] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f91ca6f1-6881-455c-bb43-98849dcf50f7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.031187] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a29b8c12-6a65-4f91-abbc-1eb1cc3ed158 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.040501] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25166aa8-3e36-45b8-95b7-c9b9b1b57819 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.057360] env[62965]: DEBUG nova.compute.provider_tree [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 601.067841] env[62965]: DEBUG nova.compute.manager [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 601.459238] env[62965]: DEBUG nova.compute.manager [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 601.491037] env[62965]: DEBUG nova.virt.hardware [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 601.491400] env[62965]: DEBUG nova.virt.hardware [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 601.491473] env[62965]: DEBUG nova.virt.hardware [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 601.492207] env[62965]: DEBUG nova.virt.hardware [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 601.492207] env[62965]: DEBUG nova.virt.hardware [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 601.492207] env[62965]: DEBUG nova.virt.hardware [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 601.492207] env[62965]: DEBUG nova.virt.hardware [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 601.492434] env[62965]: DEBUG nova.virt.hardware [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 601.492640] env[62965]: DEBUG nova.virt.hardware [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 601.492732] env[62965]: DEBUG nova.virt.hardware [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 601.492875] env[62965]: DEBUG nova.virt.hardware [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 601.493833] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8e9a4a4-d1b8-4f12-9f0f-6e21758936d5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.502879] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bcd24e3-3763-4bfe-aeda-52a1b5b1613f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.560161] env[62965]: ERROR nova.compute.manager [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3974d385-edd8-4d22-a7c2-1f79a1c724ec, please check neutron logs for more information. [ 601.560161] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 601.560161] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 601.560161] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 601.560161] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 601.560161] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 601.560161] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 601.560161] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 601.560161] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.560161] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 601.560161] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.560161] env[62965]: ERROR nova.compute.manager raise self.value [ 601.560161] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 601.560161] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 601.560161] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.560161] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 601.560719] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.560719] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 601.560719] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3974d385-edd8-4d22-a7c2-1f79a1c724ec, please check neutron logs for more information. [ 601.560719] env[62965]: ERROR nova.compute.manager [ 601.560719] env[62965]: Traceback (most recent call last): [ 601.560719] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 601.560719] env[62965]: listener.cb(fileno) [ 601.560719] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.560719] env[62965]: result = function(*args, **kwargs) [ 601.560719] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 601.560719] env[62965]: return func(*args, **kwargs) [ 601.560719] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 601.560719] env[62965]: raise e [ 601.560719] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 601.560719] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 601.560719] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 601.560719] env[62965]: created_port_ids = self._update_ports_for_instance( [ 601.560719] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 601.560719] env[62965]: with excutils.save_and_reraise_exception(): [ 601.560719] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.560719] env[62965]: self.force_reraise() [ 601.560719] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.560719] env[62965]: raise self.value [ 601.560719] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 601.560719] env[62965]: updated_port = self._update_port( [ 601.560719] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.560719] env[62965]: _ensure_no_port_binding_failure(port) [ 601.560719] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.560719] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 601.561566] env[62965]: nova.exception.PortBindingFailed: Binding failed for port 3974d385-edd8-4d22-a7c2-1f79a1c724ec, please check neutron logs for more information. [ 601.561566] env[62965]: Removing descriptor: 19 [ 601.561566] env[62965]: ERROR nova.compute.manager [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3974d385-edd8-4d22-a7c2-1f79a1c724ec, please check neutron logs for more information. [ 601.561566] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Traceback (most recent call last): [ 601.561566] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 601.561566] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] yield resources [ 601.561566] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 601.561566] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] self.driver.spawn(context, instance, image_meta, [ 601.561566] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 601.561566] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 601.561566] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 601.561566] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] vm_ref = self.build_virtual_machine(instance, [ 601.561938] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 601.561938] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] vif_infos = vmwarevif.get_vif_info(self._session, [ 601.561938] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 601.561938] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] for vif in network_info: [ 601.561938] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 601.561938] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] return self._sync_wrapper(fn, *args, **kwargs) [ 601.561938] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 601.561938] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] self.wait() [ 601.561938] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 601.561938] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] self[:] = self._gt.wait() [ 601.561938] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 601.561938] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] return self._exit_event.wait() [ 601.561938] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 601.562354] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] result = hub.switch() [ 601.562354] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 601.562354] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] return self.greenlet.switch() [ 601.562354] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.562354] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] result = function(*args, **kwargs) [ 601.562354] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 601.562354] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] return func(*args, **kwargs) [ 601.562354] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 601.562354] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] raise e [ 601.562354] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 601.562354] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] nwinfo = self.network_api.allocate_for_instance( [ 601.562354] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 601.562354] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] created_port_ids = self._update_ports_for_instance( [ 601.562751] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 601.562751] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] with excutils.save_and_reraise_exception(): [ 601.562751] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.562751] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] self.force_reraise() [ 601.562751] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.562751] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] raise self.value [ 601.562751] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 601.562751] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] updated_port = self._update_port( [ 601.562751] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.562751] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] _ensure_no_port_binding_failure(port) [ 601.562751] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.562751] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] raise exception.PortBindingFailed(port_id=port['id']) [ 601.563141] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] nova.exception.PortBindingFailed: Binding failed for port 3974d385-edd8-4d22-a7c2-1f79a1c724ec, please check neutron logs for more information. [ 601.563141] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] [ 601.563141] env[62965]: INFO nova.compute.manager [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Terminating instance [ 601.566617] env[62965]: DEBUG nova.scheduler.client.report [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 601.589787] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.072043] env[62965]: DEBUG oslo_concurrency.lockutils [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.637s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.072043] env[62965]: DEBUG nova.compute.manager [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 602.073174] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] Acquiring lock "refresh_cache-d66e557f-fbd5-4369-bdcb-14f25c3bbc7b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.073712] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] Acquired lock "refresh_cache-d66e557f-fbd5-4369-bdcb-14f25c3bbc7b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.074197] env[62965]: DEBUG nova.network.neutron [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 602.076825] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.148s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.076825] env[62965]: DEBUG nova.objects.instance [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Lazy-loading 'resources' on Instance uuid 532ae225-da05-4cbb-9bb4-c72a8451acff {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 602.254253] env[62965]: DEBUG nova.compute.manager [req-c5c02bd5-4ee2-47af-9e88-b0f81a3e8513 req-e6050f17-e5eb-4f03-8f2c-2b3bd4afebdf service nova] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Received event network-changed-3974d385-edd8-4d22-a7c2-1f79a1c724ec {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 602.254600] env[62965]: DEBUG nova.compute.manager [req-c5c02bd5-4ee2-47af-9e88-b0f81a3e8513 req-e6050f17-e5eb-4f03-8f2c-2b3bd4afebdf service nova] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Refreshing instance network info cache due to event network-changed-3974d385-edd8-4d22-a7c2-1f79a1c724ec. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 602.254600] env[62965]: DEBUG oslo_concurrency.lockutils [req-c5c02bd5-4ee2-47af-9e88-b0f81a3e8513 req-e6050f17-e5eb-4f03-8f2c-2b3bd4afebdf service nova] Acquiring lock "refresh_cache-d66e557f-fbd5-4369-bdcb-14f25c3bbc7b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.577314] env[62965]: DEBUG nova.compute.utils [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 602.582889] env[62965]: DEBUG nova.compute.manager [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 602.583997] env[62965]: DEBUG nova.network.neutron [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 602.607492] env[62965]: DEBUG nova.network.neutron [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.687734] env[62965]: DEBUG nova.network.neutron [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.689966] env[62965]: DEBUG nova.policy [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ce0e7f3429454fd18d54cbb964dfb561', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e1c6e8403409438f9ff10235cd2533bc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 603.081838] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bbde010-0326-462c-9c30-26737083502b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.089698] env[62965]: DEBUG nova.compute.manager [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 603.098158] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1034a91a-ceca-4edc-a159-f356eda4b37c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.128566] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0523e6e1-bf92-49ab-9ae9-cc7107d86798 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.137474] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfeb537a-45b7-4da6-9fd3-bc397b88ee81 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.153411] env[62965]: DEBUG nova.compute.provider_tree [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 603.192489] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] Releasing lock "refresh_cache-d66e557f-fbd5-4369-bdcb-14f25c3bbc7b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.192918] env[62965]: DEBUG nova.compute.manager [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 603.193122] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 603.193421] env[62965]: DEBUG oslo_concurrency.lockutils [req-c5c02bd5-4ee2-47af-9e88-b0f81a3e8513 req-e6050f17-e5eb-4f03-8f2c-2b3bd4afebdf service nova] Acquired lock "refresh_cache-d66e557f-fbd5-4369-bdcb-14f25c3bbc7b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.193585] env[62965]: DEBUG nova.network.neutron [req-c5c02bd5-4ee2-47af-9e88-b0f81a3e8513 req-e6050f17-e5eb-4f03-8f2c-2b3bd4afebdf service nova] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Refreshing network info cache for port 3974d385-edd8-4d22-a7c2-1f79a1c724ec {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 603.197456] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-29d559bb-9426-4c94-b2ae-d0fbe75b201f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.207059] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e934996-eab7-4ac6-9fba-e7a31f627bef {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.233592] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d66e557f-fbd5-4369-bdcb-14f25c3bbc7b could not be found. [ 603.233592] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 603.233592] env[62965]: INFO nova.compute.manager [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 603.233592] env[62965]: DEBUG oslo.service.loopingcall [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 603.233592] env[62965]: DEBUG nova.compute.manager [-] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 603.233592] env[62965]: DEBUG nova.network.neutron [-] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 603.257758] env[62965]: DEBUG nova.network.neutron [-] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 603.355718] env[62965]: DEBUG nova.network.neutron [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Successfully created port: 68e40c1c-68d7-45b1-8497-f0c9e9619416 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 603.661017] env[62965]: DEBUG nova.scheduler.client.report [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 603.720106] env[62965]: DEBUG nova.network.neutron [req-c5c02bd5-4ee2-47af-9e88-b0f81a3e8513 req-e6050f17-e5eb-4f03-8f2c-2b3bd4afebdf service nova] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 603.761626] env[62965]: DEBUG nova.network.neutron [-] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.784720] env[62965]: DEBUG nova.network.neutron [req-c5c02bd5-4ee2-47af-9e88-b0f81a3e8513 req-e6050f17-e5eb-4f03-8f2c-2b3bd4afebdf service nova] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.106259] env[62965]: DEBUG nova.compute.manager [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 604.138476] env[62965]: DEBUG nova.virt.hardware [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 604.138743] env[62965]: DEBUG nova.virt.hardware [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 604.139078] env[62965]: DEBUG nova.virt.hardware [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 604.139165] env[62965]: DEBUG nova.virt.hardware [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 604.139262] env[62965]: DEBUG nova.virt.hardware [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 604.139441] env[62965]: DEBUG nova.virt.hardware [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 604.139654] env[62965]: DEBUG nova.virt.hardware [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 604.139839] env[62965]: DEBUG nova.virt.hardware [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 604.141336] env[62965]: DEBUG nova.virt.hardware [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 604.141840] env[62965]: DEBUG nova.virt.hardware [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 604.141840] env[62965]: DEBUG nova.virt.hardware [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 604.142670] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6d50abb-47c5-4a7f-a597-6d7e70eabbdd {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.156961] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33d5f44b-43fa-4786-abd5-8e93b9d76622 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.173389] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.097s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.176977] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.885s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.217087] env[62965]: INFO nova.scheduler.client.report [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Deleted allocations for instance 532ae225-da05-4cbb-9bb4-c72a8451acff [ 604.266741] env[62965]: INFO nova.compute.manager [-] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Took 1.03 seconds to deallocate network for instance. [ 604.271842] env[62965]: DEBUG nova.compute.claims [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 604.272132] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.295945] env[62965]: DEBUG oslo_concurrency.lockutils [req-c5c02bd5-4ee2-47af-9e88-b0f81a3e8513 req-e6050f17-e5eb-4f03-8f2c-2b3bd4afebdf service nova] Releasing lock "refresh_cache-d66e557f-fbd5-4369-bdcb-14f25c3bbc7b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 604.296211] env[62965]: DEBUG nova.compute.manager [req-c5c02bd5-4ee2-47af-9e88-b0f81a3e8513 req-e6050f17-e5eb-4f03-8f2c-2b3bd4afebdf service nova] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Received event network-vif-deleted-3974d385-edd8-4d22-a7c2-1f79a1c724ec {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 604.728243] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7beda060-bbd7-40ec-9fc6-a321688af42a tempest-ServerDiagnosticsV248Test-587608252 tempest-ServerDiagnosticsV248Test-587608252-project-member] Lock "532ae225-da05-4cbb-9bb4-c72a8451acff" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.318s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.115764] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c682929f-2501-4e67-afa2-d163ea172766 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.126736] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe79a116-0786-4b7d-bd6d-e9533c366797 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.161883] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d60a7a18-be16-4644-aedb-af00fd27f0c3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.171584] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7452bbd5-b838-422b-991d-68d820a8d8b1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.188655] env[62965]: DEBUG nova.compute.provider_tree [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 605.317231] env[62965]: DEBUG nova.compute.manager [req-212ae137-ef61-4e49-8b83-7eccbf9e7e6f req-27e08cc4-db9d-45b0-90e1-3df697090d62 service nova] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Received event network-changed-68e40c1c-68d7-45b1-8497-f0c9e9619416 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 605.319597] env[62965]: DEBUG nova.compute.manager [req-212ae137-ef61-4e49-8b83-7eccbf9e7e6f req-27e08cc4-db9d-45b0-90e1-3df697090d62 service nova] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Refreshing instance network info cache due to event network-changed-68e40c1c-68d7-45b1-8497-f0c9e9619416. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 605.320615] env[62965]: DEBUG oslo_concurrency.lockutils [req-212ae137-ef61-4e49-8b83-7eccbf9e7e6f req-27e08cc4-db9d-45b0-90e1-3df697090d62 service nova] Acquiring lock "refresh_cache-2931c8b9-ed5f-4994-a4d5-50e1369d67e9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.320615] env[62965]: DEBUG oslo_concurrency.lockutils [req-212ae137-ef61-4e49-8b83-7eccbf9e7e6f req-27e08cc4-db9d-45b0-90e1-3df697090d62 service nova] Acquired lock "refresh_cache-2931c8b9-ed5f-4994-a4d5-50e1369d67e9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.320615] env[62965]: DEBUG nova.network.neutron [req-212ae137-ef61-4e49-8b83-7eccbf9e7e6f req-27e08cc4-db9d-45b0-90e1-3df697090d62 service nova] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Refreshing network info cache for port 68e40c1c-68d7-45b1-8497-f0c9e9619416 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 605.388785] env[62965]: ERROR nova.compute.manager [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 68e40c1c-68d7-45b1-8497-f0c9e9619416, please check neutron logs for more information. [ 605.388785] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 605.388785] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 605.388785] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 605.388785] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 605.388785] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 605.388785] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 605.388785] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 605.388785] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.388785] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 605.388785] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.388785] env[62965]: ERROR nova.compute.manager raise self.value [ 605.388785] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 605.388785] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 605.388785] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.388785] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 605.389571] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.389571] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 605.389571] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 68e40c1c-68d7-45b1-8497-f0c9e9619416, please check neutron logs for more information. [ 605.389571] env[62965]: ERROR nova.compute.manager [ 605.389571] env[62965]: Traceback (most recent call last): [ 605.389571] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 605.389571] env[62965]: listener.cb(fileno) [ 605.389571] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 605.389571] env[62965]: result = function(*args, **kwargs) [ 605.389571] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 605.389571] env[62965]: return func(*args, **kwargs) [ 605.389571] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 605.389571] env[62965]: raise e [ 605.389571] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 605.389571] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 605.389571] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 605.389571] env[62965]: created_port_ids = self._update_ports_for_instance( [ 605.389571] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 605.389571] env[62965]: with excutils.save_and_reraise_exception(): [ 605.389571] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.389571] env[62965]: self.force_reraise() [ 605.389571] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.389571] env[62965]: raise self.value [ 605.389571] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 605.389571] env[62965]: updated_port = self._update_port( [ 605.389571] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.389571] env[62965]: _ensure_no_port_binding_failure(port) [ 605.389571] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.389571] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 605.390505] env[62965]: nova.exception.PortBindingFailed: Binding failed for port 68e40c1c-68d7-45b1-8497-f0c9e9619416, please check neutron logs for more information. [ 605.390505] env[62965]: Removing descriptor: 17 [ 605.390505] env[62965]: ERROR nova.compute.manager [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 68e40c1c-68d7-45b1-8497-f0c9e9619416, please check neutron logs for more information. [ 605.390505] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Traceback (most recent call last): [ 605.390505] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 605.390505] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] yield resources [ 605.390505] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 605.390505] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] self.driver.spawn(context, instance, image_meta, [ 605.390505] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 605.390505] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 605.390505] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 605.390505] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] vm_ref = self.build_virtual_machine(instance, [ 605.390895] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 605.390895] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] vif_infos = vmwarevif.get_vif_info(self._session, [ 605.390895] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 605.390895] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] for vif in network_info: [ 605.390895] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 605.390895] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] return self._sync_wrapper(fn, *args, **kwargs) [ 605.390895] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 605.390895] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] self.wait() [ 605.390895] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 605.390895] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] self[:] = self._gt.wait() [ 605.390895] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 605.390895] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] return self._exit_event.wait() [ 605.390895] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 605.391306] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] result = hub.switch() [ 605.391306] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 605.391306] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] return self.greenlet.switch() [ 605.391306] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 605.391306] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] result = function(*args, **kwargs) [ 605.391306] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 605.391306] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] return func(*args, **kwargs) [ 605.391306] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 605.391306] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] raise e [ 605.391306] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 605.391306] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] nwinfo = self.network_api.allocate_for_instance( [ 605.391306] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 605.391306] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] created_port_ids = self._update_ports_for_instance( [ 605.391718] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 605.391718] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] with excutils.save_and_reraise_exception(): [ 605.391718] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.391718] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] self.force_reraise() [ 605.391718] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.391718] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] raise self.value [ 605.391718] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 605.391718] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] updated_port = self._update_port( [ 605.391718] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.391718] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] _ensure_no_port_binding_failure(port) [ 605.391718] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.391718] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] raise exception.PortBindingFailed(port_id=port['id']) [ 605.392141] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] nova.exception.PortBindingFailed: Binding failed for port 68e40c1c-68d7-45b1-8497-f0c9e9619416, please check neutron logs for more information. [ 605.392141] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] [ 605.392141] env[62965]: INFO nova.compute.manager [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Terminating instance [ 605.693154] env[62965]: DEBUG nova.scheduler.client.report [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 605.866450] env[62965]: DEBUG nova.network.neutron [req-212ae137-ef61-4e49-8b83-7eccbf9e7e6f req-27e08cc4-db9d-45b0-90e1-3df697090d62 service nova] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 605.898961] env[62965]: DEBUG oslo_concurrency.lockutils [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquiring lock "refresh_cache-2931c8b9-ed5f-4994-a4d5-50e1369d67e9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.966285] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] Acquiring lock "8342b9df-ebfd-403b-a0b8-1fdd6ebe2add" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.966582] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] Lock "8342b9df-ebfd-403b-a0b8-1fdd6ebe2add" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.013318] env[62965]: DEBUG nova.network.neutron [req-212ae137-ef61-4e49-8b83-7eccbf9e7e6f req-27e08cc4-db9d-45b0-90e1-3df697090d62 service nova] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.200545] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.021s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.200545] env[62965]: ERROR nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9add2eb2-422d-493e-ac3d-acf980ff6c90, please check neutron logs for more information. [ 606.200545] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Traceback (most recent call last): [ 606.200545] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 606.200545] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] self.driver.spawn(context, instance, image_meta, [ 606.200545] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 606.200545] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 606.200545] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 606.200545] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] vm_ref = self.build_virtual_machine(instance, [ 606.200919] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 606.200919] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] vif_infos = vmwarevif.get_vif_info(self._session, [ 606.200919] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 606.200919] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] for vif in network_info: [ 606.200919] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 606.200919] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] return self._sync_wrapper(fn, *args, **kwargs) [ 606.200919] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 606.200919] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] self.wait() [ 606.200919] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 606.200919] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] self[:] = self._gt.wait() [ 606.200919] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 606.200919] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] return self._exit_event.wait() [ 606.200919] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 606.201394] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] result = hub.switch() [ 606.201394] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 606.201394] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] return self.greenlet.switch() [ 606.201394] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 606.201394] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] result = function(*args, **kwargs) [ 606.201394] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 606.201394] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] return func(*args, **kwargs) [ 606.201394] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 606.201394] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] raise e [ 606.201394] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 606.201394] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] nwinfo = self.network_api.allocate_for_instance( [ 606.201394] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 606.201394] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] created_port_ids = self._update_ports_for_instance( [ 606.201789] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 606.201789] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] with excutils.save_and_reraise_exception(): [ 606.201789] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.201789] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] self.force_reraise() [ 606.201789] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.201789] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] raise self.value [ 606.201789] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 606.201789] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] updated_port = self._update_port( [ 606.201789] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.201789] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] _ensure_no_port_binding_failure(port) [ 606.201789] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.201789] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] raise exception.PortBindingFailed(port_id=port['id']) [ 606.202164] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] nova.exception.PortBindingFailed: Binding failed for port 9add2eb2-422d-493e-ac3d-acf980ff6c90, please check neutron logs for more information. [ 606.202164] env[62965]: ERROR nova.compute.manager [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] [ 606.202164] env[62965]: DEBUG nova.compute.utils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Binding failed for port 9add2eb2-422d-493e-ac3d-acf980ff6c90, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 606.202164] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.484s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.203869] env[62965]: INFO nova.compute.claims [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 606.206531] env[62965]: DEBUG nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Build of instance 2870d8cb-8c19-454e-8a6e-1675eba8ddd4 was re-scheduled: Binding failed for port 9add2eb2-422d-493e-ac3d-acf980ff6c90, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 606.206983] env[62965]: DEBUG nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 606.207218] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Acquiring lock "refresh_cache-2870d8cb-8c19-454e-8a6e-1675eba8ddd4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 606.207373] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Acquired lock "refresh_cache-2870d8cb-8c19-454e-8a6e-1675eba8ddd4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.207585] env[62965]: DEBUG nova.network.neutron [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 606.515881] env[62965]: DEBUG oslo_concurrency.lockutils [req-212ae137-ef61-4e49-8b83-7eccbf9e7e6f req-27e08cc4-db9d-45b0-90e1-3df697090d62 service nova] Releasing lock "refresh_cache-2931c8b9-ed5f-4994-a4d5-50e1369d67e9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.516303] env[62965]: DEBUG oslo_concurrency.lockutils [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquired lock "refresh_cache-2931c8b9-ed5f-4994-a4d5-50e1369d67e9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.516486] env[62965]: DEBUG nova.network.neutron [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 606.731203] env[62965]: DEBUG nova.network.neutron [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.884422] env[62965]: DEBUG nova.network.neutron [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.037792] env[62965]: DEBUG nova.network.neutron [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 607.159933] env[62965]: DEBUG nova.network.neutron [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.384039] env[62965]: DEBUG nova.compute.manager [req-c65737bb-1cb1-42e1-b3aa-fd98f83723ad req-dfe204dc-31f0-4d49-964a-5616cdb2c00c service nova] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Received event network-vif-deleted-68e40c1c-68d7-45b1-8497-f0c9e9619416 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 607.391489] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Releasing lock "refresh_cache-2870d8cb-8c19-454e-8a6e-1675eba8ddd4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 607.391766] env[62965]: DEBUG nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 607.391880] env[62965]: DEBUG nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 607.392058] env[62965]: DEBUG nova.network.neutron [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 607.412308] env[62965]: DEBUG nova.network.neutron [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 607.666341] env[62965]: DEBUG oslo_concurrency.lockutils [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Releasing lock "refresh_cache-2931c8b9-ed5f-4994-a4d5-50e1369d67e9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 607.666612] env[62965]: DEBUG nova.compute.manager [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 607.666794] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 607.667199] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1da9cc12-bc76-4b2a-8707-f8e025bf0366 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.670954] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08dac1b1-4dce-46ed-884a-db9a91d22d3a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.679714] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cad3d468-916e-4075-8be9-f96d93ef1f01 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.686256] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97ad3b32-afff-4202-b513-f85dcc36416b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.739828] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc53ba8f-2e7f-4ccb-b990-73dccbe07e66 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.743412] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2931c8b9-ed5f-4994-a4d5-50e1369d67e9 could not be found. [ 607.743624] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 607.743797] env[62965]: INFO nova.compute.manager [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Took 0.08 seconds to destroy the instance on the hypervisor. [ 607.744417] env[62965]: DEBUG oslo.service.loopingcall [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 607.744417] env[62965]: DEBUG nova.compute.manager [-] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 607.744417] env[62965]: DEBUG nova.network.neutron [-] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 607.750820] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a887da3-1910-49ca-94ce-b087e4ec85d9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.764401] env[62965]: DEBUG nova.compute.provider_tree [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 607.768030] env[62965]: DEBUG nova.network.neutron [-] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 607.916483] env[62965]: DEBUG nova.network.neutron [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.172780] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] Acquiring lock "cd978d56-ba80-44a9-9a02-3a1254066571" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.173085] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] Lock "cd978d56-ba80-44a9-9a02-3a1254066571" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.267782] env[62965]: DEBUG nova.scheduler.client.report [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 608.270976] env[62965]: DEBUG nova.network.neutron [-] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.419631] env[62965]: INFO nova.compute.manager [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] [instance: 2870d8cb-8c19-454e-8a6e-1675eba8ddd4] Took 1.03 seconds to deallocate network for instance. [ 608.774201] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.572s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.774736] env[62965]: DEBUG nova.compute.manager [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 608.777353] env[62965]: INFO nova.compute.manager [-] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Took 1.03 seconds to deallocate network for instance. [ 608.777886] env[62965]: DEBUG oslo_concurrency.lockutils [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.318s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.779411] env[62965]: INFO nova.compute.claims [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 608.785632] env[62965]: DEBUG nova.compute.claims [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 608.785800] env[62965]: DEBUG oslo_concurrency.lockutils [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.279326] env[62965]: DEBUG nova.compute.utils [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 609.280766] env[62965]: DEBUG nova.compute.manager [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 609.280937] env[62965]: DEBUG nova.network.neutron [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 609.364322] env[62965]: DEBUG nova.policy [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8cc49c20f2d549f587cca42df25962f8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd7298fbffd9d491e9bb59440f5889521', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 609.460153] env[62965]: INFO nova.scheduler.client.report [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Deleted allocations for instance 2870d8cb-8c19-454e-8a6e-1675eba8ddd4 [ 609.740411] env[62965]: DEBUG nova.network.neutron [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Successfully created port: 0e5fdb64-ae19-4373-87ba-620ccd0339f2 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 609.783720] env[62965]: DEBUG nova.compute.manager [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 609.973188] env[62965]: DEBUG oslo_concurrency.lockutils [None req-62b5408f-7865-41a3-b729-2e00c31fcfda tempest-ListServersNegativeTestJSON-199871936 tempest-ListServersNegativeTestJSON-199871936-project-member] Lock "2870d8cb-8c19-454e-8a6e-1675eba8ddd4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 64.235s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 610.297614] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-104555d0-353b-4a8b-bd75-e9895a4b1a90 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.305831] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1bd5261-d364-40fd-841c-12e31b3e2c64 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.335697] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-603b8c90-bb82-4f1e-b16f-d25b16fb8bbd {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.344617] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4246a014-749c-48eb-86b8-3b110704cb51 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.360151] env[62965]: DEBUG nova.compute.provider_tree [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 610.475790] env[62965]: DEBUG nova.compute.manager [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 610.794745] env[62965]: DEBUG nova.compute.manager [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 610.829603] env[62965]: DEBUG nova.virt.hardware [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 610.829834] env[62965]: DEBUG nova.virt.hardware [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 610.829983] env[62965]: DEBUG nova.virt.hardware [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 610.830169] env[62965]: DEBUG nova.virt.hardware [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 610.830306] env[62965]: DEBUG nova.virt.hardware [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 610.830480] env[62965]: DEBUG nova.virt.hardware [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 610.830700] env[62965]: DEBUG nova.virt.hardware [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 610.830856] env[62965]: DEBUG nova.virt.hardware [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 610.831035] env[62965]: DEBUG nova.virt.hardware [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 610.831197] env[62965]: DEBUG nova.virt.hardware [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 610.831363] env[62965]: DEBUG nova.virt.hardware [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 610.832285] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17520530-1292-46ea-ac33-abc9041f5324 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.840393] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f600118e-4223-473d-90b0-83a033b0cba8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.863750] env[62965]: DEBUG nova.scheduler.client.report [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 610.918125] env[62965]: DEBUG nova.compute.manager [req-3fedf380-861d-4266-8c13-5ee9bc0d9018 req-f98e7ea8-1808-44e9-a373-cd4cc9cb9031 service nova] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Received event network-changed-0e5fdb64-ae19-4373-87ba-620ccd0339f2 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 610.918125] env[62965]: DEBUG nova.compute.manager [req-3fedf380-861d-4266-8c13-5ee9bc0d9018 req-f98e7ea8-1808-44e9-a373-cd4cc9cb9031 service nova] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Refreshing instance network info cache due to event network-changed-0e5fdb64-ae19-4373-87ba-620ccd0339f2. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 610.918125] env[62965]: DEBUG oslo_concurrency.lockutils [req-3fedf380-861d-4266-8c13-5ee9bc0d9018 req-f98e7ea8-1808-44e9-a373-cd4cc9cb9031 service nova] Acquiring lock "refresh_cache-f619dd08-0cdf-486f-8f13-707b254a8593" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 610.918125] env[62965]: DEBUG oslo_concurrency.lockutils [req-3fedf380-861d-4266-8c13-5ee9bc0d9018 req-f98e7ea8-1808-44e9-a373-cd4cc9cb9031 service nova] Acquired lock "refresh_cache-f619dd08-0cdf-486f-8f13-707b254a8593" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.918739] env[62965]: DEBUG nova.network.neutron [req-3fedf380-861d-4266-8c13-5ee9bc0d9018 req-f98e7ea8-1808-44e9-a373-cd4cc9cb9031 service nova] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Refreshing network info cache for port 0e5fdb64-ae19-4373-87ba-620ccd0339f2 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 611.000899] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.088346] env[62965]: ERROR nova.compute.manager [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0e5fdb64-ae19-4373-87ba-620ccd0339f2, please check neutron logs for more information. [ 611.088346] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 611.088346] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 611.088346] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 611.088346] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.088346] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 611.088346] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.088346] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 611.088346] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.088346] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 611.088346] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.088346] env[62965]: ERROR nova.compute.manager raise self.value [ 611.088346] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.088346] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 611.088346] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.088346] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 611.088849] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.088849] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 611.088849] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0e5fdb64-ae19-4373-87ba-620ccd0339f2, please check neutron logs for more information. [ 611.088849] env[62965]: ERROR nova.compute.manager [ 611.088849] env[62965]: Traceback (most recent call last): [ 611.088849] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 611.088849] env[62965]: listener.cb(fileno) [ 611.088849] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.088849] env[62965]: result = function(*args, **kwargs) [ 611.088849] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 611.088849] env[62965]: return func(*args, **kwargs) [ 611.088849] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 611.088849] env[62965]: raise e [ 611.088849] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 611.088849] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 611.088849] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.088849] env[62965]: created_port_ids = self._update_ports_for_instance( [ 611.088849] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.088849] env[62965]: with excutils.save_and_reraise_exception(): [ 611.088849] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.088849] env[62965]: self.force_reraise() [ 611.088849] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.088849] env[62965]: raise self.value [ 611.088849] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.088849] env[62965]: updated_port = self._update_port( [ 611.088849] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.088849] env[62965]: _ensure_no_port_binding_failure(port) [ 611.088849] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.088849] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 611.089764] env[62965]: nova.exception.PortBindingFailed: Binding failed for port 0e5fdb64-ae19-4373-87ba-620ccd0339f2, please check neutron logs for more information. [ 611.089764] env[62965]: Removing descriptor: 17 [ 611.089764] env[62965]: ERROR nova.compute.manager [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0e5fdb64-ae19-4373-87ba-620ccd0339f2, please check neutron logs for more information. [ 611.089764] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Traceback (most recent call last): [ 611.089764] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 611.089764] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] yield resources [ 611.089764] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 611.089764] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] self.driver.spawn(context, instance, image_meta, [ 611.089764] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 611.089764] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] self._vmops.spawn(context, instance, image_meta, injected_files, [ 611.089764] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 611.089764] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] vm_ref = self.build_virtual_machine(instance, [ 611.090171] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 611.090171] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] vif_infos = vmwarevif.get_vif_info(self._session, [ 611.090171] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 611.090171] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] for vif in network_info: [ 611.090171] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 611.090171] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] return self._sync_wrapper(fn, *args, **kwargs) [ 611.090171] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 611.090171] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] self.wait() [ 611.090171] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 611.090171] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] self[:] = self._gt.wait() [ 611.090171] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 611.090171] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] return self._exit_event.wait() [ 611.090171] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 611.090571] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] result = hub.switch() [ 611.090571] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 611.090571] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] return self.greenlet.switch() [ 611.090571] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.090571] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] result = function(*args, **kwargs) [ 611.090571] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 611.090571] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] return func(*args, **kwargs) [ 611.090571] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 611.090571] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] raise e [ 611.090571] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 611.090571] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] nwinfo = self.network_api.allocate_for_instance( [ 611.090571] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.090571] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] created_port_ids = self._update_ports_for_instance( [ 611.090989] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.090989] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] with excutils.save_and_reraise_exception(): [ 611.090989] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.090989] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] self.force_reraise() [ 611.090989] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.090989] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] raise self.value [ 611.090989] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.090989] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] updated_port = self._update_port( [ 611.090989] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.090989] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] _ensure_no_port_binding_failure(port) [ 611.090989] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.090989] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] raise exception.PortBindingFailed(port_id=port['id']) [ 611.091363] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] nova.exception.PortBindingFailed: Binding failed for port 0e5fdb64-ae19-4373-87ba-620ccd0339f2, please check neutron logs for more information. [ 611.091363] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] [ 611.091363] env[62965]: INFO nova.compute.manager [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Terminating instance [ 611.368378] env[62965]: DEBUG oslo_concurrency.lockutils [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.590s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 611.369096] env[62965]: DEBUG nova.compute.manager [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 611.371596] env[62965]: DEBUG oslo_concurrency.lockutils [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.751s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.373083] env[62965]: INFO nova.compute.claims [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 611.499808] env[62965]: DEBUG nova.network.neutron [req-3fedf380-861d-4266-8c13-5ee9bc0d9018 req-f98e7ea8-1808-44e9-a373-cd4cc9cb9031 service nova] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.594869] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Acquiring lock "refresh_cache-f619dd08-0cdf-486f-8f13-707b254a8593" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.620071] env[62965]: DEBUG nova.network.neutron [req-3fedf380-861d-4266-8c13-5ee9bc0d9018 req-f98e7ea8-1808-44e9-a373-cd4cc9cb9031 service nova] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.879169] env[62965]: DEBUG nova.compute.utils [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 611.881389] env[62965]: DEBUG nova.compute.manager [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 611.881697] env[62965]: DEBUG nova.network.neutron [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 611.941246] env[62965]: DEBUG nova.policy [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '540ab66ac47a472f8d93409ff1706777', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b4c59e1d6b5b4ad0bb0e46a72dfc5a2c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 612.127717] env[62965]: DEBUG oslo_concurrency.lockutils [req-3fedf380-861d-4266-8c13-5ee9bc0d9018 req-f98e7ea8-1808-44e9-a373-cd4cc9cb9031 service nova] Releasing lock "refresh_cache-f619dd08-0cdf-486f-8f13-707b254a8593" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.127717] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Acquired lock "refresh_cache-f619dd08-0cdf-486f-8f13-707b254a8593" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.127717] env[62965]: DEBUG nova.network.neutron [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 612.386086] env[62965]: DEBUG nova.network.neutron [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Successfully created port: ce89fee2-26e8-4f78-9de6-148d957ccb38 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 612.386086] env[62965]: DEBUG nova.compute.manager [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 612.660836] env[62965]: DEBUG nova.network.neutron [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.757158] env[62965]: DEBUG nova.network.neutron [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.942220] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e292029-69df-433f-ac4b-1fcc49a9d9c8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.951130] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b101d4f-0b44-4855-bd92-2042fa4d1e7e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.984946] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8f7c2da-acf2-45fe-8f59-88e074c2948d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.993069] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0646ba95-6447-4bf7-9adf-f2c6e19e5e51 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.006668] env[62965]: DEBUG nova.compute.provider_tree [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 613.067086] env[62965]: DEBUG nova.compute.manager [req-c50b51b0-7162-4e18-96d5-47b1b157c14d req-2b5d7062-a7ea-4d85-833f-124ade9c6261 service nova] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Received event network-vif-deleted-0e5fdb64-ae19-4373-87ba-620ccd0339f2 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 613.260362] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Releasing lock "refresh_cache-f619dd08-0cdf-486f-8f13-707b254a8593" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.260777] env[62965]: DEBUG nova.compute.manager [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 613.261023] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 613.261336] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f5f5c27f-beda-4ef1-bf6c-321326f0f9be {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.273748] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ed104ff-0b12-4afb-8431-658460c1f3e8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.305010] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f619dd08-0cdf-486f-8f13-707b254a8593 could not be found. [ 613.305727] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 613.305727] env[62965]: INFO nova.compute.manager [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Took 0.04 seconds to destroy the instance on the hypervisor. [ 613.305727] env[62965]: DEBUG oslo.service.loopingcall [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 613.305872] env[62965]: DEBUG nova.compute.manager [-] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 613.306117] env[62965]: DEBUG nova.network.neutron [-] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 613.327613] env[62965]: DEBUG nova.network.neutron [-] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.400310] env[62965]: DEBUG nova.compute.manager [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 613.431506] env[62965]: DEBUG nova.virt.hardware [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 613.431617] env[62965]: DEBUG nova.virt.hardware [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 613.431710] env[62965]: DEBUG nova.virt.hardware [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 613.431886] env[62965]: DEBUG nova.virt.hardware [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 613.432045] env[62965]: DEBUG nova.virt.hardware [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 613.432196] env[62965]: DEBUG nova.virt.hardware [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 613.432392] env[62965]: DEBUG nova.virt.hardware [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 613.432545] env[62965]: DEBUG nova.virt.hardware [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 613.432706] env[62965]: DEBUG nova.virt.hardware [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 613.432860] env[62965]: DEBUG nova.virt.hardware [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 613.433038] env[62965]: DEBUG nova.virt.hardware [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 613.434152] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddab067b-1f29-4718-a832-c54105923fa0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.442381] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff9c41a1-d1b9-4289-9898-f8aa3b8eabb6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.509549] env[62965]: DEBUG nova.scheduler.client.report [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 613.672861] env[62965]: ERROR nova.compute.manager [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ce89fee2-26e8-4f78-9de6-148d957ccb38, please check neutron logs for more information. [ 613.672861] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 613.672861] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 613.672861] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 613.672861] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 613.672861] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 613.672861] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 613.672861] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 613.672861] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.672861] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 613.672861] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.672861] env[62965]: ERROR nova.compute.manager raise self.value [ 613.672861] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 613.672861] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 613.672861] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.672861] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 613.673377] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.673377] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 613.673377] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ce89fee2-26e8-4f78-9de6-148d957ccb38, please check neutron logs for more information. [ 613.673377] env[62965]: ERROR nova.compute.manager [ 613.673377] env[62965]: Traceback (most recent call last): [ 613.673377] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 613.673377] env[62965]: listener.cb(fileno) [ 613.673377] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 613.673377] env[62965]: result = function(*args, **kwargs) [ 613.673377] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 613.673377] env[62965]: return func(*args, **kwargs) [ 613.673377] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 613.673377] env[62965]: raise e [ 613.673377] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 613.673377] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 613.673377] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 613.673377] env[62965]: created_port_ids = self._update_ports_for_instance( [ 613.673377] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 613.673377] env[62965]: with excutils.save_and_reraise_exception(): [ 613.673377] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.673377] env[62965]: self.force_reraise() [ 613.673377] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.673377] env[62965]: raise self.value [ 613.673377] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 613.673377] env[62965]: updated_port = self._update_port( [ 613.673377] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.673377] env[62965]: _ensure_no_port_binding_failure(port) [ 613.673377] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.673377] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 613.673934] env[62965]: nova.exception.PortBindingFailed: Binding failed for port ce89fee2-26e8-4f78-9de6-148d957ccb38, please check neutron logs for more information. [ 613.673934] env[62965]: Removing descriptor: 17 [ 613.673934] env[62965]: ERROR nova.compute.manager [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ce89fee2-26e8-4f78-9de6-148d957ccb38, please check neutron logs for more information. [ 613.673934] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Traceback (most recent call last): [ 613.673934] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 613.673934] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] yield resources [ 613.673934] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 613.673934] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] self.driver.spawn(context, instance, image_meta, [ 613.673934] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 613.673934] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] self._vmops.spawn(context, instance, image_meta, injected_files, [ 613.673934] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 613.673934] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] vm_ref = self.build_virtual_machine(instance, [ 613.674196] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 613.674196] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] vif_infos = vmwarevif.get_vif_info(self._session, [ 613.674196] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 613.674196] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] for vif in network_info: [ 613.674196] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 613.674196] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] return self._sync_wrapper(fn, *args, **kwargs) [ 613.674196] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 613.674196] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] self.wait() [ 613.674196] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 613.674196] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] self[:] = self._gt.wait() [ 613.674196] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 613.674196] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] return self._exit_event.wait() [ 613.674196] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 613.674519] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] result = hub.switch() [ 613.674519] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 613.674519] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] return self.greenlet.switch() [ 613.674519] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 613.674519] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] result = function(*args, **kwargs) [ 613.674519] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 613.674519] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] return func(*args, **kwargs) [ 613.674519] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 613.674519] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] raise e [ 613.674519] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 613.674519] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] nwinfo = self.network_api.allocate_for_instance( [ 613.674519] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 613.674519] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] created_port_ids = self._update_ports_for_instance( [ 613.674794] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 613.674794] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] with excutils.save_and_reraise_exception(): [ 613.674794] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.674794] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] self.force_reraise() [ 613.674794] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.674794] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] raise self.value [ 613.674794] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 613.674794] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] updated_port = self._update_port( [ 613.674794] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.674794] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] _ensure_no_port_binding_failure(port) [ 613.674794] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.674794] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] raise exception.PortBindingFailed(port_id=port['id']) [ 613.675047] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] nova.exception.PortBindingFailed: Binding failed for port ce89fee2-26e8-4f78-9de6-148d957ccb38, please check neutron logs for more information. [ 613.675047] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] [ 613.675047] env[62965]: INFO nova.compute.manager [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Terminating instance [ 613.830549] env[62965]: DEBUG nova.network.neutron [-] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.015567] env[62965]: DEBUG oslo_concurrency.lockutils [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.644s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.016109] env[62965]: DEBUG nova.compute.manager [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 614.022165] env[62965]: DEBUG oslo_concurrency.lockutils [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.291s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.024400] env[62965]: INFO nova.compute.claims [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 614.177696] env[62965]: DEBUG oslo_concurrency.lockutils [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] Acquiring lock "refresh_cache-61702d54-59d2-4d13-9eed-59b052b7fe74" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.177924] env[62965]: DEBUG oslo_concurrency.lockutils [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] Acquired lock "refresh_cache-61702d54-59d2-4d13-9eed-59b052b7fe74" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.178242] env[62965]: DEBUG nova.network.neutron [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 614.332887] env[62965]: INFO nova.compute.manager [-] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Took 1.03 seconds to deallocate network for instance. [ 614.335340] env[62965]: DEBUG nova.compute.claims [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 614.335516] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.527605] env[62965]: DEBUG nova.compute.utils [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 614.529550] env[62965]: DEBUG nova.compute.manager [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 614.529744] env[62965]: DEBUG nova.network.neutron [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 614.569187] env[62965]: DEBUG nova.policy [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0b988aec43614cef89bcf346d34ab8f7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e65c7a070146461c8e7adaba09aeda35', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 614.705781] env[62965]: DEBUG nova.network.neutron [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 614.833466] env[62965]: DEBUG nova.network.neutron [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.885675] env[62965]: DEBUG nova.network.neutron [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Successfully created port: adf3e41d-5e61-406e-b35a-fb511e5bfcaa {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 615.030657] env[62965]: DEBUG nova.compute.manager [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 615.092576] env[62965]: DEBUG nova.compute.manager [req-940fb3ce-3305-49cf-b16b-4b7d9c070a2a req-f09d4b5f-a429-4eca-b550-a431a78b16a0 service nova] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Received event network-changed-ce89fee2-26e8-4f78-9de6-148d957ccb38 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 615.092853] env[62965]: DEBUG nova.compute.manager [req-940fb3ce-3305-49cf-b16b-4b7d9c070a2a req-f09d4b5f-a429-4eca-b550-a431a78b16a0 service nova] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Refreshing instance network info cache due to event network-changed-ce89fee2-26e8-4f78-9de6-148d957ccb38. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 615.092952] env[62965]: DEBUG oslo_concurrency.lockutils [req-940fb3ce-3305-49cf-b16b-4b7d9c070a2a req-f09d4b5f-a429-4eca-b550-a431a78b16a0 service nova] Acquiring lock "refresh_cache-61702d54-59d2-4d13-9eed-59b052b7fe74" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 615.338581] env[62965]: DEBUG oslo_concurrency.lockutils [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] Releasing lock "refresh_cache-61702d54-59d2-4d13-9eed-59b052b7fe74" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.339139] env[62965]: DEBUG nova.compute.manager [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 615.339451] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 615.339910] env[62965]: DEBUG oslo_concurrency.lockutils [req-940fb3ce-3305-49cf-b16b-4b7d9c070a2a req-f09d4b5f-a429-4eca-b550-a431a78b16a0 service nova] Acquired lock "refresh_cache-61702d54-59d2-4d13-9eed-59b052b7fe74" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.340213] env[62965]: DEBUG nova.network.neutron [req-940fb3ce-3305-49cf-b16b-4b7d9c070a2a req-f09d4b5f-a429-4eca-b550-a431a78b16a0 service nova] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Refreshing network info cache for port ce89fee2-26e8-4f78-9de6-148d957ccb38 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 615.341318] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-547198ef-abce-4020-9a56-3bedb6f5ff3f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.356431] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bd99f14-d694-4b87-9879-7327152fead2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.380958] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 61702d54-59d2-4d13-9eed-59b052b7fe74 could not be found. [ 615.381191] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 615.381368] env[62965]: INFO nova.compute.manager [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Took 0.04 seconds to destroy the instance on the hypervisor. [ 615.381621] env[62965]: DEBUG oslo.service.loopingcall [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 615.383858] env[62965]: DEBUG nova.compute.manager [-] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 615.383959] env[62965]: DEBUG nova.network.neutron [-] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 615.405766] env[62965]: DEBUG nova.network.neutron [-] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 615.532454] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14fffe4e-af9c-4c1c-ba1d-6afbfc345164 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.542982] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e2bf77f-db82-4d29-bffa-89d12d124810 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.575820] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36106b99-a5ba-4646-8555-c88afc66ff34 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.584689] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82c6725e-d163-48f7-a862-1912ca7f4e39 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.600147] env[62965]: DEBUG nova.compute.provider_tree [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 615.879345] env[62965]: DEBUG nova.network.neutron [req-940fb3ce-3305-49cf-b16b-4b7d9c070a2a req-f09d4b5f-a429-4eca-b550-a431a78b16a0 service nova] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 615.908759] env[62965]: DEBUG nova.network.neutron [-] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.968475] env[62965]: DEBUG nova.network.neutron [req-940fb3ce-3305-49cf-b16b-4b7d9c070a2a req-f09d4b5f-a429-4eca-b550-a431a78b16a0 service nova] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.041050] env[62965]: DEBUG nova.compute.manager [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 616.082118] env[62965]: DEBUG nova.virt.hardware [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:31:19Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='316658043',id=25,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-782873223',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 616.082118] env[62965]: DEBUG nova.virt.hardware [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 616.082118] env[62965]: DEBUG nova.virt.hardware [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 616.082486] env[62965]: DEBUG nova.virt.hardware [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 616.083107] env[62965]: DEBUG nova.virt.hardware [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 616.083654] env[62965]: DEBUG nova.virt.hardware [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 616.083958] env[62965]: DEBUG nova.virt.hardware [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 616.084288] env[62965]: DEBUG nova.virt.hardware [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 616.084660] env[62965]: DEBUG nova.virt.hardware [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 616.085183] env[62965]: DEBUG nova.virt.hardware [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 616.085470] env[62965]: DEBUG nova.virt.hardware [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 616.086807] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e44135d8-4b9a-493e-8c32-f2a2b6dd7b89 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.096094] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38ff54e5-e0fd-45d3-91b5-830af6c36f95 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.112542] env[62965]: DEBUG nova.scheduler.client.report [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 616.414772] env[62965]: INFO nova.compute.manager [-] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Took 1.03 seconds to deallocate network for instance. [ 616.415461] env[62965]: DEBUG nova.compute.claims [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 616.415807] env[62965]: DEBUG oslo_concurrency.lockutils [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.424095] env[62965]: ERROR nova.compute.manager [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port adf3e41d-5e61-406e-b35a-fb511e5bfcaa, please check neutron logs for more information. [ 616.424095] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 616.424095] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 616.424095] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 616.424095] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 616.424095] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 616.424095] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 616.424095] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 616.424095] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 616.424095] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 616.424095] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 616.424095] env[62965]: ERROR nova.compute.manager raise self.value [ 616.424095] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 616.424095] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 616.424095] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 616.424095] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 616.424435] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 616.424435] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 616.424435] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port adf3e41d-5e61-406e-b35a-fb511e5bfcaa, please check neutron logs for more information. [ 616.424435] env[62965]: ERROR nova.compute.manager [ 616.424435] env[62965]: Traceback (most recent call last): [ 616.424435] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 616.424435] env[62965]: listener.cb(fileno) [ 616.424435] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 616.424435] env[62965]: result = function(*args, **kwargs) [ 616.424435] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 616.424435] env[62965]: return func(*args, **kwargs) [ 616.424435] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 616.424435] env[62965]: raise e [ 616.424435] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 616.424435] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 616.424435] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 616.424435] env[62965]: created_port_ids = self._update_ports_for_instance( [ 616.424435] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 616.424435] env[62965]: with excutils.save_and_reraise_exception(): [ 616.424435] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 616.424435] env[62965]: self.force_reraise() [ 616.424435] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 616.424435] env[62965]: raise self.value [ 616.424435] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 616.424435] env[62965]: updated_port = self._update_port( [ 616.424435] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 616.424435] env[62965]: _ensure_no_port_binding_failure(port) [ 616.424435] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 616.424435] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 616.425021] env[62965]: nova.exception.PortBindingFailed: Binding failed for port adf3e41d-5e61-406e-b35a-fb511e5bfcaa, please check neutron logs for more information. [ 616.425021] env[62965]: Removing descriptor: 17 [ 616.425021] env[62965]: ERROR nova.compute.manager [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port adf3e41d-5e61-406e-b35a-fb511e5bfcaa, please check neutron logs for more information. [ 616.425021] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Traceback (most recent call last): [ 616.425021] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 616.425021] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] yield resources [ 616.425021] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 616.425021] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] self.driver.spawn(context, instance, image_meta, [ 616.425021] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 616.425021] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] self._vmops.spawn(context, instance, image_meta, injected_files, [ 616.425021] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 616.425021] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] vm_ref = self.build_virtual_machine(instance, [ 616.425266] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 616.425266] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] vif_infos = vmwarevif.get_vif_info(self._session, [ 616.425266] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 616.425266] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] for vif in network_info: [ 616.425266] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 616.425266] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] return self._sync_wrapper(fn, *args, **kwargs) [ 616.425266] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 616.425266] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] self.wait() [ 616.425266] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 616.425266] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] self[:] = self._gt.wait() [ 616.425266] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 616.425266] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] return self._exit_event.wait() [ 616.425266] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 616.425548] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] result = hub.switch() [ 616.425548] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 616.425548] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] return self.greenlet.switch() [ 616.425548] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 616.425548] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] result = function(*args, **kwargs) [ 616.425548] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 616.425548] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] return func(*args, **kwargs) [ 616.425548] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 616.425548] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] raise e [ 616.425548] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 616.425548] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] nwinfo = self.network_api.allocate_for_instance( [ 616.425548] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 616.425548] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] created_port_ids = self._update_ports_for_instance( [ 616.425811] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 616.425811] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] with excutils.save_and_reraise_exception(): [ 616.425811] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 616.425811] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] self.force_reraise() [ 616.425811] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 616.425811] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] raise self.value [ 616.425811] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 616.425811] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] updated_port = self._update_port( [ 616.425811] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 616.425811] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] _ensure_no_port_binding_failure(port) [ 616.425811] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 616.425811] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] raise exception.PortBindingFailed(port_id=port['id']) [ 616.426058] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] nova.exception.PortBindingFailed: Binding failed for port adf3e41d-5e61-406e-b35a-fb511e5bfcaa, please check neutron logs for more information. [ 616.426058] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] [ 616.426058] env[62965]: INFO nova.compute.manager [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Terminating instance [ 616.473022] env[62965]: DEBUG oslo_concurrency.lockutils [req-940fb3ce-3305-49cf-b16b-4b7d9c070a2a req-f09d4b5f-a429-4eca-b550-a431a78b16a0 service nova] Releasing lock "refresh_cache-61702d54-59d2-4d13-9eed-59b052b7fe74" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 616.473292] env[62965]: DEBUG nova.compute.manager [req-940fb3ce-3305-49cf-b16b-4b7d9c070a2a req-f09d4b5f-a429-4eca-b550-a431a78b16a0 service nova] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Received event network-vif-deleted-ce89fee2-26e8-4f78-9de6-148d957ccb38 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 616.618905] env[62965]: DEBUG oslo_concurrency.lockutils [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.597s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.619434] env[62965]: DEBUG nova.compute.manager [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 616.622220] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.784s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.625857] env[62965]: INFO nova.compute.claims [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 616.930558] env[62965]: DEBUG oslo_concurrency.lockutils [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] Acquiring lock "refresh_cache-59187ca2-0482-4603-8689-542d2bf5d023" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 616.930750] env[62965]: DEBUG oslo_concurrency.lockutils [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] Acquired lock "refresh_cache-59187ca2-0482-4603-8689-542d2bf5d023" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.930927] env[62965]: DEBUG nova.network.neutron [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 617.129462] env[62965]: DEBUG nova.compute.utils [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 617.136248] env[62965]: DEBUG nova.compute.manager [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 617.136248] env[62965]: DEBUG nova.network.neutron [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 617.194419] env[62965]: DEBUG nova.compute.manager [req-38a17aca-a4dd-47f5-ac3a-443a2e1834ee req-b278085a-9444-49a7-849f-d31888816c4d service nova] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Received event network-changed-adf3e41d-5e61-406e-b35a-fb511e5bfcaa {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 617.194835] env[62965]: DEBUG nova.compute.manager [req-38a17aca-a4dd-47f5-ac3a-443a2e1834ee req-b278085a-9444-49a7-849f-d31888816c4d service nova] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Refreshing instance network info cache due to event network-changed-adf3e41d-5e61-406e-b35a-fb511e5bfcaa. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 617.195116] env[62965]: DEBUG oslo_concurrency.lockutils [req-38a17aca-a4dd-47f5-ac3a-443a2e1834ee req-b278085a-9444-49a7-849f-d31888816c4d service nova] Acquiring lock "refresh_cache-59187ca2-0482-4603-8689-542d2bf5d023" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 617.226485] env[62965]: DEBUG nova.policy [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1fe38180500e4610b505c48479ee4cd9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '14b716820b354f2d82b357bfa5c2afa5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 617.452375] env[62965]: DEBUG nova.network.neutron [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 617.581186] env[62965]: DEBUG nova.network.neutron [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.614366] env[62965]: DEBUG nova.network.neutron [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Successfully created port: b5e4287e-3c4b-4299-a138-4be56eede9c4 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 617.636788] env[62965]: DEBUG nova.compute.manager [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 618.084141] env[62965]: DEBUG oslo_concurrency.lockutils [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] Releasing lock "refresh_cache-59187ca2-0482-4603-8689-542d2bf5d023" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 618.084141] env[62965]: DEBUG nova.compute.manager [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 618.085158] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 618.085548] env[62965]: DEBUG oslo_concurrency.lockutils [req-38a17aca-a4dd-47f5-ac3a-443a2e1834ee req-b278085a-9444-49a7-849f-d31888816c4d service nova] Acquired lock "refresh_cache-59187ca2-0482-4603-8689-542d2bf5d023" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.085548] env[62965]: DEBUG nova.network.neutron [req-38a17aca-a4dd-47f5-ac3a-443a2e1834ee req-b278085a-9444-49a7-849f-d31888816c4d service nova] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Refreshing network info cache for port adf3e41d-5e61-406e-b35a-fb511e5bfcaa {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 618.087566] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d76a6316-d46d-40ef-b29a-0068258a2656 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.100965] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-582651f1-d6c6-4620-a527-b446ae7f427f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.133488] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 59187ca2-0482-4603-8689-542d2bf5d023 could not be found. [ 618.133769] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 618.133956] env[62965]: INFO nova.compute.manager [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Took 0.05 seconds to destroy the instance on the hypervisor. [ 618.134221] env[62965]: DEBUG oslo.service.loopingcall [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 618.136704] env[62965]: DEBUG nova.compute.manager [-] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 618.136807] env[62965]: DEBUG nova.network.neutron [-] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 618.154315] env[62965]: DEBUG nova.network.neutron [-] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 618.169921] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-653d3ca0-ee58-43b0-8e3a-c3e7fcec10b3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.179564] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3235ac7-61e9-41cb-b3e1-adfe1c35511f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.213192] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dd3217f-9b3b-4d1f-9cef-cfb436929326 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.222715] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-600a8b77-5998-4838-979d-569c64481bfe {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.240320] env[62965]: DEBUG nova.compute.provider_tree [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 618.606317] env[62965]: DEBUG nova.network.neutron [req-38a17aca-a4dd-47f5-ac3a-443a2e1834ee req-b278085a-9444-49a7-849f-d31888816c4d service nova] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 618.658049] env[62965]: DEBUG nova.network.neutron [-] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.675628] env[62965]: DEBUG nova.compute.manager [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 618.684858] env[62965]: DEBUG nova.network.neutron [req-38a17aca-a4dd-47f5-ac3a-443a2e1834ee req-b278085a-9444-49a7-849f-d31888816c4d service nova] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.702535] env[62965]: DEBUG nova.virt.hardware [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 618.702774] env[62965]: DEBUG nova.virt.hardware [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 618.702930] env[62965]: DEBUG nova.virt.hardware [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 618.703124] env[62965]: DEBUG nova.virt.hardware [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 618.703266] env[62965]: DEBUG nova.virt.hardware [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 618.703406] env[62965]: DEBUG nova.virt.hardware [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 618.703601] env[62965]: DEBUG nova.virt.hardware [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 618.703753] env[62965]: DEBUG nova.virt.hardware [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 618.703914] env[62965]: DEBUG nova.virt.hardware [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 618.704170] env[62965]: DEBUG nova.virt.hardware [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 618.704364] env[62965]: DEBUG nova.virt.hardware [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 618.705404] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22132035-5404-40bc-bf3f-2239cdcedd1a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.712919] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb2520d6-65c7-4e36-95da-1653e739a27b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.744977] env[62965]: DEBUG nova.scheduler.client.report [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 618.900992] env[62965]: ERROR nova.compute.manager [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b5e4287e-3c4b-4299-a138-4be56eede9c4, please check neutron logs for more information. [ 618.900992] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 618.900992] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 618.900992] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 618.900992] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 618.900992] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 618.900992] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 618.900992] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 618.900992] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.900992] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 618.900992] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.900992] env[62965]: ERROR nova.compute.manager raise self.value [ 618.900992] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 618.900992] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 618.900992] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.900992] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 618.901462] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.901462] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 618.901462] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b5e4287e-3c4b-4299-a138-4be56eede9c4, please check neutron logs for more information. [ 618.901462] env[62965]: ERROR nova.compute.manager [ 618.901462] env[62965]: Traceback (most recent call last): [ 618.901462] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 618.901462] env[62965]: listener.cb(fileno) [ 618.901462] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.901462] env[62965]: result = function(*args, **kwargs) [ 618.901462] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 618.901462] env[62965]: return func(*args, **kwargs) [ 618.901462] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 618.901462] env[62965]: raise e [ 618.901462] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 618.901462] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 618.901462] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 618.901462] env[62965]: created_port_ids = self._update_ports_for_instance( [ 618.901462] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 618.901462] env[62965]: with excutils.save_and_reraise_exception(): [ 618.901462] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.901462] env[62965]: self.force_reraise() [ 618.901462] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.901462] env[62965]: raise self.value [ 618.901462] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 618.901462] env[62965]: updated_port = self._update_port( [ 618.901462] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.901462] env[62965]: _ensure_no_port_binding_failure(port) [ 618.901462] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.901462] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 618.902196] env[62965]: nova.exception.PortBindingFailed: Binding failed for port b5e4287e-3c4b-4299-a138-4be56eede9c4, please check neutron logs for more information. [ 618.902196] env[62965]: Removing descriptor: 17 [ 618.902196] env[62965]: ERROR nova.compute.manager [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b5e4287e-3c4b-4299-a138-4be56eede9c4, please check neutron logs for more information. [ 618.902196] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Traceback (most recent call last): [ 618.902196] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 618.902196] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] yield resources [ 618.902196] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 618.902196] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] self.driver.spawn(context, instance, image_meta, [ 618.902196] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 618.902196] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 618.902196] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 618.902196] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] vm_ref = self.build_virtual_machine(instance, [ 618.902511] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 618.902511] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] vif_infos = vmwarevif.get_vif_info(self._session, [ 618.902511] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 618.902511] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] for vif in network_info: [ 618.902511] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 618.902511] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] return self._sync_wrapper(fn, *args, **kwargs) [ 618.902511] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 618.902511] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] self.wait() [ 618.902511] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 618.902511] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] self[:] = self._gt.wait() [ 618.902511] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 618.902511] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] return self._exit_event.wait() [ 618.902511] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 618.902843] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] result = hub.switch() [ 618.902843] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 618.902843] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] return self.greenlet.switch() [ 618.902843] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.902843] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] result = function(*args, **kwargs) [ 618.902843] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 618.902843] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] return func(*args, **kwargs) [ 618.902843] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 618.902843] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] raise e [ 618.902843] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 618.902843] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] nwinfo = self.network_api.allocate_for_instance( [ 618.902843] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 618.902843] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] created_port_ids = self._update_ports_for_instance( [ 618.903180] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 618.903180] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] with excutils.save_and_reraise_exception(): [ 618.903180] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.903180] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] self.force_reraise() [ 618.903180] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.903180] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] raise self.value [ 618.903180] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 618.903180] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] updated_port = self._update_port( [ 618.903180] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.903180] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] _ensure_no_port_binding_failure(port) [ 618.903180] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.903180] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] raise exception.PortBindingFailed(port_id=port['id']) [ 618.903480] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] nova.exception.PortBindingFailed: Binding failed for port b5e4287e-3c4b-4299-a138-4be56eede9c4, please check neutron logs for more information. [ 618.903480] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] [ 618.903480] env[62965]: INFO nova.compute.manager [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Terminating instance [ 619.162924] env[62965]: INFO nova.compute.manager [-] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Took 1.02 seconds to deallocate network for instance. [ 619.164022] env[62965]: DEBUG nova.compute.claims [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 619.164215] env[62965]: DEBUG oslo_concurrency.lockutils [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.187426] env[62965]: DEBUG oslo_concurrency.lockutils [req-38a17aca-a4dd-47f5-ac3a-443a2e1834ee req-b278085a-9444-49a7-849f-d31888816c4d service nova] Releasing lock "refresh_cache-59187ca2-0482-4603-8689-542d2bf5d023" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.187749] env[62965]: DEBUG nova.compute.manager [req-38a17aca-a4dd-47f5-ac3a-443a2e1834ee req-b278085a-9444-49a7-849f-d31888816c4d service nova] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Received event network-vif-deleted-adf3e41d-5e61-406e-b35a-fb511e5bfcaa {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 619.221791] env[62965]: DEBUG nova.compute.manager [req-3842d6a1-99ea-4b65-b733-e1ec96641d31 req-e5feb506-0c32-4f3a-8b54-3c001a744132 service nova] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Received event network-changed-b5e4287e-3c4b-4299-a138-4be56eede9c4 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 619.222033] env[62965]: DEBUG nova.compute.manager [req-3842d6a1-99ea-4b65-b733-e1ec96641d31 req-e5feb506-0c32-4f3a-8b54-3c001a744132 service nova] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Refreshing instance network info cache due to event network-changed-b5e4287e-3c4b-4299-a138-4be56eede9c4. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 619.222202] env[62965]: DEBUG oslo_concurrency.lockutils [req-3842d6a1-99ea-4b65-b733-e1ec96641d31 req-e5feb506-0c32-4f3a-8b54-3c001a744132 service nova] Acquiring lock "refresh_cache-53519d3e-34d2-410f-ac19-69a015e407b9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.222343] env[62965]: DEBUG oslo_concurrency.lockutils [req-3842d6a1-99ea-4b65-b733-e1ec96641d31 req-e5feb506-0c32-4f3a-8b54-3c001a744132 service nova] Acquired lock "refresh_cache-53519d3e-34d2-410f-ac19-69a015e407b9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.222497] env[62965]: DEBUG nova.network.neutron [req-3842d6a1-99ea-4b65-b733-e1ec96641d31 req-e5feb506-0c32-4f3a-8b54-3c001a744132 service nova] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Refreshing network info cache for port b5e4287e-3c4b-4299-a138-4be56eede9c4 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 619.249533] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.627s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.250034] env[62965]: DEBUG nova.compute.manager [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 619.254338] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.285s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.406097] env[62965]: DEBUG oslo_concurrency.lockutils [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] Acquiring lock "refresh_cache-53519d3e-34d2-410f-ac19-69a015e407b9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.739749] env[62965]: DEBUG nova.network.neutron [req-3842d6a1-99ea-4b65-b733-e1ec96641d31 req-e5feb506-0c32-4f3a-8b54-3c001a744132 service nova] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 619.758261] env[62965]: DEBUG nova.compute.utils [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 619.762506] env[62965]: DEBUG nova.compute.manager [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 619.762735] env[62965]: DEBUG nova.network.neutron [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 619.803033] env[62965]: DEBUG nova.policy [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dc2af663375640c88a35792dca53a198', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '93413fa776be4e7cb4111f16721a8688', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 619.843303] env[62965]: DEBUG nova.network.neutron [req-3842d6a1-99ea-4b65-b733-e1ec96641d31 req-e5feb506-0c32-4f3a-8b54-3c001a744132 service nova] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.072360] env[62965]: DEBUG nova.network.neutron [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Successfully created port: 8dc9700a-cf9b-4c75-b4a1-e1c34262d46e {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 620.150229] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ff9eccc-2230-4e15-9380-74f9b35e15ee {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.157510] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb87f5a6-39f8-409a-9ecd-69587bd3eba9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.191013] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d170f4c-8a85-4d8d-bc96-a01027e8c3f6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.198600] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a017458c-ac92-48ae-a8b4-71403db976a9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.212279] env[62965]: DEBUG nova.compute.provider_tree [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.265563] env[62965]: DEBUG nova.compute.manager [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 620.347554] env[62965]: DEBUG oslo_concurrency.lockutils [req-3842d6a1-99ea-4b65-b733-e1ec96641d31 req-e5feb506-0c32-4f3a-8b54-3c001a744132 service nova] Releasing lock "refresh_cache-53519d3e-34d2-410f-ac19-69a015e407b9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.347554] env[62965]: DEBUG nova.compute.manager [req-3842d6a1-99ea-4b65-b733-e1ec96641d31 req-e5feb506-0c32-4f3a-8b54-3c001a744132 service nova] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Received event network-vif-deleted-b5e4287e-3c4b-4299-a138-4be56eede9c4 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 620.347554] env[62965]: DEBUG oslo_concurrency.lockutils [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] Acquired lock "refresh_cache-53519d3e-34d2-410f-ac19-69a015e407b9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 620.347554] env[62965]: DEBUG nova.network.neutron [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 620.714949] env[62965]: DEBUG nova.scheduler.client.report [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 620.723756] env[62965]: DEBUG nova.compute.manager [req-ae655b2e-536a-475e-b027-e6b0af8da8fa req-3f6bccb3-0175-4f31-8056-637cdeb6a8aa service nova] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Received event network-changed-8dc9700a-cf9b-4c75-b4a1-e1c34262d46e {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 620.723945] env[62965]: DEBUG nova.compute.manager [req-ae655b2e-536a-475e-b027-e6b0af8da8fa req-3f6bccb3-0175-4f31-8056-637cdeb6a8aa service nova] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Refreshing instance network info cache due to event network-changed-8dc9700a-cf9b-4c75-b4a1-e1c34262d46e. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 620.724164] env[62965]: DEBUG oslo_concurrency.lockutils [req-ae655b2e-536a-475e-b027-e6b0af8da8fa req-3f6bccb3-0175-4f31-8056-637cdeb6a8aa service nova] Acquiring lock "refresh_cache-f594c50e-25b3-4e85-840c-7dfe0994a2c2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 620.724296] env[62965]: DEBUG oslo_concurrency.lockutils [req-ae655b2e-536a-475e-b027-e6b0af8da8fa req-3f6bccb3-0175-4f31-8056-637cdeb6a8aa service nova] Acquired lock "refresh_cache-f594c50e-25b3-4e85-840c-7dfe0994a2c2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 620.724444] env[62965]: DEBUG nova.network.neutron [req-ae655b2e-536a-475e-b027-e6b0af8da8fa req-3f6bccb3-0175-4f31-8056-637cdeb6a8aa service nova] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Refreshing network info cache for port 8dc9700a-cf9b-4c75-b4a1-e1c34262d46e {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 620.868245] env[62965]: DEBUG nova.network.neutron [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.901111] env[62965]: ERROR nova.compute.manager [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8dc9700a-cf9b-4c75-b4a1-e1c34262d46e, please check neutron logs for more information. [ 620.901111] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 620.901111] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 620.901111] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 620.901111] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 620.901111] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 620.901111] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 620.901111] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 620.901111] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 620.901111] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 620.901111] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 620.901111] env[62965]: ERROR nova.compute.manager raise self.value [ 620.901111] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 620.901111] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 620.901111] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 620.901111] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 620.901545] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 620.901545] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 620.901545] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8dc9700a-cf9b-4c75-b4a1-e1c34262d46e, please check neutron logs for more information. [ 620.901545] env[62965]: ERROR nova.compute.manager [ 620.901545] env[62965]: Traceback (most recent call last): [ 620.901545] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 620.901545] env[62965]: listener.cb(fileno) [ 620.901545] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 620.901545] env[62965]: result = function(*args, **kwargs) [ 620.901545] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 620.901545] env[62965]: return func(*args, **kwargs) [ 620.901545] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 620.901545] env[62965]: raise e [ 620.901545] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 620.901545] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 620.901545] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 620.901545] env[62965]: created_port_ids = self._update_ports_for_instance( [ 620.901545] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 620.901545] env[62965]: with excutils.save_and_reraise_exception(): [ 620.901545] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 620.901545] env[62965]: self.force_reraise() [ 620.901545] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 620.901545] env[62965]: raise self.value [ 620.901545] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 620.901545] env[62965]: updated_port = self._update_port( [ 620.901545] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 620.901545] env[62965]: _ensure_no_port_binding_failure(port) [ 620.901545] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 620.901545] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 620.902283] env[62965]: nova.exception.PortBindingFailed: Binding failed for port 8dc9700a-cf9b-4c75-b4a1-e1c34262d46e, please check neutron logs for more information. [ 620.902283] env[62965]: Removing descriptor: 19 [ 621.109110] env[62965]: DEBUG nova.network.neutron [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.220450] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.967s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.221080] env[62965]: ERROR nova.compute.manager [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3211a0c8-0637-4db2-befc-3acd54e53435, please check neutron logs for more information. [ 621.221080] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Traceback (most recent call last): [ 621.221080] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 621.221080] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] self.driver.spawn(context, instance, image_meta, [ 621.221080] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 621.221080] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 621.221080] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 621.221080] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] vm_ref = self.build_virtual_machine(instance, [ 621.221080] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 621.221080] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] vif_infos = vmwarevif.get_vif_info(self._session, [ 621.221080] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 621.221420] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] for vif in network_info: [ 621.221420] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 621.221420] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] return self._sync_wrapper(fn, *args, **kwargs) [ 621.221420] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 621.221420] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] self.wait() [ 621.221420] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 621.221420] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] self[:] = self._gt.wait() [ 621.221420] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 621.221420] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] return self._exit_event.wait() [ 621.221420] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 621.221420] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] result = hub.switch() [ 621.221420] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 621.221420] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] return self.greenlet.switch() [ 621.221755] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 621.221755] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] result = function(*args, **kwargs) [ 621.221755] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 621.221755] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] return func(*args, **kwargs) [ 621.221755] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 621.221755] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] raise e [ 621.221755] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 621.221755] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] nwinfo = self.network_api.allocate_for_instance( [ 621.221755] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 621.221755] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] created_port_ids = self._update_ports_for_instance( [ 621.221755] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 621.221755] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] with excutils.save_and_reraise_exception(): [ 621.221755] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.222162] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] self.force_reraise() [ 621.222162] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.222162] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] raise self.value [ 621.222162] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 621.222162] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] updated_port = self._update_port( [ 621.222162] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.222162] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] _ensure_no_port_binding_failure(port) [ 621.222162] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.222162] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] raise exception.PortBindingFailed(port_id=port['id']) [ 621.222162] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] nova.exception.PortBindingFailed: Binding failed for port 3211a0c8-0637-4db2-befc-3acd54e53435, please check neutron logs for more information. [ 621.222162] env[62965]: ERROR nova.compute.manager [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] [ 621.222447] env[62965]: DEBUG nova.compute.utils [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Binding failed for port 3211a0c8-0637-4db2-befc-3acd54e53435, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 621.222987] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.633s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.224496] env[62965]: INFO nova.compute.claims [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 621.226985] env[62965]: DEBUG nova.compute.manager [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Build of instance 647a3ba2-8b84-45f9-9d28-130eb0521d1b was re-scheduled: Binding failed for port 3211a0c8-0637-4db2-befc-3acd54e53435, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 621.227437] env[62965]: DEBUG nova.compute.manager [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 621.227684] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Acquiring lock "refresh_cache-647a3ba2-8b84-45f9-9d28-130eb0521d1b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 621.227926] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Acquired lock "refresh_cache-647a3ba2-8b84-45f9-9d28-130eb0521d1b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.227995] env[62965]: DEBUG nova.network.neutron [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 621.243237] env[62965]: DEBUG nova.network.neutron [req-ae655b2e-536a-475e-b027-e6b0af8da8fa req-3f6bccb3-0175-4f31-8056-637cdeb6a8aa service nova] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 621.271878] env[62965]: DEBUG nova.compute.manager [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 621.299506] env[62965]: DEBUG nova.virt.hardware [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 621.299751] env[62965]: DEBUG nova.virt.hardware [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 621.299902] env[62965]: DEBUG nova.virt.hardware [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 621.300089] env[62965]: DEBUG nova.virt.hardware [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 621.300232] env[62965]: DEBUG nova.virt.hardware [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 621.300369] env[62965]: DEBUG nova.virt.hardware [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 621.300568] env[62965]: DEBUG nova.virt.hardware [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 621.300748] env[62965]: DEBUG nova.virt.hardware [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 621.300938] env[62965]: DEBUG nova.virt.hardware [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 621.301111] env[62965]: DEBUG nova.virt.hardware [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 621.301282] env[62965]: DEBUG nova.virt.hardware [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 621.302429] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7bd6536-6298-4a8d-b36e-3a446376fa94 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.312575] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23465219-a734-4e18-a092-c211a73cda4d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.326986] env[62965]: ERROR nova.compute.manager [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8dc9700a-cf9b-4c75-b4a1-e1c34262d46e, please check neutron logs for more information. [ 621.326986] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Traceback (most recent call last): [ 621.326986] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 621.326986] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] yield resources [ 621.326986] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 621.326986] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] self.driver.spawn(context, instance, image_meta, [ 621.326986] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 621.326986] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 621.326986] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 621.326986] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] vm_ref = self.build_virtual_machine(instance, [ 621.326986] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 621.327323] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] vif_infos = vmwarevif.get_vif_info(self._session, [ 621.327323] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 621.327323] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] for vif in network_info: [ 621.327323] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 621.327323] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] return self._sync_wrapper(fn, *args, **kwargs) [ 621.327323] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 621.327323] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] self.wait() [ 621.327323] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 621.327323] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] self[:] = self._gt.wait() [ 621.327323] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 621.327323] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] return self._exit_event.wait() [ 621.327323] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 621.327323] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] current.throw(*self._exc) [ 621.327598] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 621.327598] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] result = function(*args, **kwargs) [ 621.327598] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 621.327598] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] return func(*args, **kwargs) [ 621.327598] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 621.327598] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] raise e [ 621.327598] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 621.327598] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] nwinfo = self.network_api.allocate_for_instance( [ 621.327598] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 621.327598] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] created_port_ids = self._update_ports_for_instance( [ 621.327598] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 621.327598] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] with excutils.save_and_reraise_exception(): [ 621.327598] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.327889] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] self.force_reraise() [ 621.327889] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.327889] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] raise self.value [ 621.327889] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 621.327889] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] updated_port = self._update_port( [ 621.327889] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.327889] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] _ensure_no_port_binding_failure(port) [ 621.327889] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.327889] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] raise exception.PortBindingFailed(port_id=port['id']) [ 621.327889] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] nova.exception.PortBindingFailed: Binding failed for port 8dc9700a-cf9b-4c75-b4a1-e1c34262d46e, please check neutron logs for more information. [ 621.327889] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] [ 621.327889] env[62965]: INFO nova.compute.manager [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Terminating instance [ 621.376771] env[62965]: DEBUG nova.network.neutron [req-ae655b2e-536a-475e-b027-e6b0af8da8fa req-3f6bccb3-0175-4f31-8056-637cdeb6a8aa service nova] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.613281] env[62965]: DEBUG oslo_concurrency.lockutils [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] Releasing lock "refresh_cache-53519d3e-34d2-410f-ac19-69a015e407b9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 621.613281] env[62965]: DEBUG nova.compute.manager [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 621.613407] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 621.613691] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d2dcdb10-70e1-44c4-88c0-46a7426fd5f7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.622162] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b4e56a4-e7f4-4284-ac54-c0e1e463ba50 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.643963] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 53519d3e-34d2-410f-ac19-69a015e407b9 could not be found. [ 621.644205] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 621.644385] env[62965]: INFO nova.compute.manager [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Took 0.03 seconds to destroy the instance on the hypervisor. [ 621.644619] env[62965]: DEBUG oslo.service.loopingcall [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 621.644835] env[62965]: DEBUG nova.compute.manager [-] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 621.644932] env[62965]: DEBUG nova.network.neutron [-] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 621.659129] env[62965]: DEBUG nova.network.neutron [-] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 621.746159] env[62965]: DEBUG nova.network.neutron [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 621.813096] env[62965]: DEBUG nova.network.neutron [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.831414] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] Acquiring lock "refresh_cache-f594c50e-25b3-4e85-840c-7dfe0994a2c2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 621.880771] env[62965]: DEBUG oslo_concurrency.lockutils [req-ae655b2e-536a-475e-b027-e6b0af8da8fa req-3f6bccb3-0175-4f31-8056-637cdeb6a8aa service nova] Releasing lock "refresh_cache-f594c50e-25b3-4e85-840c-7dfe0994a2c2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 621.881192] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] Acquired lock "refresh_cache-f594c50e-25b3-4e85-840c-7dfe0994a2c2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.881371] env[62965]: DEBUG nova.network.neutron [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 622.161457] env[62965]: DEBUG nova.network.neutron [-] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.315393] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Releasing lock "refresh_cache-647a3ba2-8b84-45f9-9d28-130eb0521d1b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 622.315626] env[62965]: DEBUG nova.compute.manager [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 622.315807] env[62965]: DEBUG nova.compute.manager [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 622.315970] env[62965]: DEBUG nova.network.neutron [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 622.342445] env[62965]: DEBUG nova.network.neutron [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.404146] env[62965]: DEBUG nova.network.neutron [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.510411] env[62965]: DEBUG nova.network.neutron [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.603534] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc8e2265-f36a-461e-827a-d5ec0a2eaeac {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.611141] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7df47e0a-4803-4358-9728-5051bb27ac46 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.643017] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-521fe45e-af35-4ed7-bec3-adce76ba13b8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.650190] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f5dfcf6-5a94-47a5-9724-0bbd29c9abcd {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.663587] env[62965]: INFO nova.compute.manager [-] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Took 1.02 seconds to deallocate network for instance. [ 622.664033] env[62965]: DEBUG nova.compute.provider_tree [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 622.666911] env[62965]: DEBUG nova.compute.claims [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 622.667107] env[62965]: DEBUG oslo_concurrency.lockutils [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.750317] env[62965]: DEBUG nova.compute.manager [req-da2c47b9-8523-45ab-93f2-9858e01a43d2 req-ade33726-1cdd-4e1d-ac9a-eaa7ba3745e2 service nova] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Received event network-vif-deleted-8dc9700a-cf9b-4c75-b4a1-e1c34262d46e {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 622.845887] env[62965]: DEBUG nova.network.neutron [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.016862] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] Releasing lock "refresh_cache-f594c50e-25b3-4e85-840c-7dfe0994a2c2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.017304] env[62965]: DEBUG nova.compute.manager [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 623.017494] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 623.017830] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-155cdfa1-3019-4875-b5c2-88e05b0d58c1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.026295] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb299d03-6921-4524-9fd7-95be7dc2be19 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.047391] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f594c50e-25b3-4e85-840c-7dfe0994a2c2 could not be found. [ 623.047584] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 623.047780] env[62965]: INFO nova.compute.manager [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Took 0.03 seconds to destroy the instance on the hypervisor. [ 623.048016] env[62965]: DEBUG oslo.service.loopingcall [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 623.048224] env[62965]: DEBUG nova.compute.manager [-] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 623.048319] env[62965]: DEBUG nova.network.neutron [-] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 623.062438] env[62965]: DEBUG nova.network.neutron [-] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 623.168319] env[62965]: DEBUG nova.scheduler.client.report [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 623.348649] env[62965]: INFO nova.compute.manager [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: 647a3ba2-8b84-45f9-9d28-130eb0521d1b] Took 1.03 seconds to deallocate network for instance. [ 623.564553] env[62965]: DEBUG nova.network.neutron [-] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.673880] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.450s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.673880] env[62965]: DEBUG nova.compute.manager [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 623.676360] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.404s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.067337] env[62965]: INFO nova.compute.manager [-] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Took 1.02 seconds to deallocate network for instance. [ 624.069659] env[62965]: DEBUG nova.compute.claims [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 624.069831] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.181167] env[62965]: DEBUG nova.compute.utils [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 624.186061] env[62965]: DEBUG nova.compute.manager [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 624.186061] env[62965]: DEBUG nova.network.neutron [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 624.250028] env[62965]: DEBUG nova.policy [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b586afcd74b34c36b03dd6f35f2e962b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e1013f4d0e864d3f9630f65f2f9a643e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 624.379807] env[62965]: INFO nova.scheduler.client.report [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Deleted allocations for instance 647a3ba2-8b84-45f9-9d28-130eb0521d1b [ 624.586167] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43dd2e90-5f13-4e6a-9ecf-8c06b21c2692 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.598311] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5efd0ecc-37d5-4362-bb28-169847bb95b7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.631870] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52bee5b0-72c3-4a6e-9cb1-7a35a35fb842 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.640413] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34ba7f59-02b6-4943-a266-fac3cda90c4b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.645987] env[62965]: DEBUG nova.network.neutron [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Successfully created port: b6c21631-dcbf-4d90-ad9b-fa7ddd03e087 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 624.657530] env[62965]: DEBUG nova.compute.provider_tree [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 624.686305] env[62965]: DEBUG nova.compute.manager [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 624.891908] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b644d330-e48f-4c67-a5ff-411759023b6c tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Lock "647a3ba2-8b84-45f9-9d28-130eb0521d1b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 75.360s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.161515] env[62965]: DEBUG nova.scheduler.client.report [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 625.397971] env[62965]: DEBUG nova.compute.manager [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 625.667261] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.991s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.668162] env[62965]: ERROR nova.compute.manager [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3974d385-edd8-4d22-a7c2-1f79a1c724ec, please check neutron logs for more information. [ 625.668162] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Traceback (most recent call last): [ 625.668162] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 625.668162] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] self.driver.spawn(context, instance, image_meta, [ 625.668162] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 625.668162] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 625.668162] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 625.668162] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] vm_ref = self.build_virtual_machine(instance, [ 625.668162] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 625.668162] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] vif_infos = vmwarevif.get_vif_info(self._session, [ 625.668162] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 625.668637] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] for vif in network_info: [ 625.668637] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 625.668637] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] return self._sync_wrapper(fn, *args, **kwargs) [ 625.668637] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 625.668637] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] self.wait() [ 625.668637] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 625.668637] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] self[:] = self._gt.wait() [ 625.668637] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 625.668637] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] return self._exit_event.wait() [ 625.668637] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 625.668637] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] result = hub.switch() [ 625.668637] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 625.668637] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] return self.greenlet.switch() [ 625.670314] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 625.670314] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] result = function(*args, **kwargs) [ 625.670314] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 625.670314] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] return func(*args, **kwargs) [ 625.670314] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 625.670314] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] raise e [ 625.670314] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 625.670314] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] nwinfo = self.network_api.allocate_for_instance( [ 625.670314] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 625.670314] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] created_port_ids = self._update_ports_for_instance( [ 625.670314] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 625.670314] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] with excutils.save_and_reraise_exception(): [ 625.670314] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.670691] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] self.force_reraise() [ 625.670691] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.670691] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] raise self.value [ 625.670691] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 625.670691] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] updated_port = self._update_port( [ 625.670691] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.670691] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] _ensure_no_port_binding_failure(port) [ 625.670691] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.670691] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] raise exception.PortBindingFailed(port_id=port['id']) [ 625.670691] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] nova.exception.PortBindingFailed: Binding failed for port 3974d385-edd8-4d22-a7c2-1f79a1c724ec, please check neutron logs for more information. [ 625.670691] env[62965]: ERROR nova.compute.manager [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] [ 625.670942] env[62965]: DEBUG nova.compute.utils [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Binding failed for port 3974d385-edd8-4d22-a7c2-1f79a1c724ec, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 625.670942] env[62965]: DEBUG nova.compute.manager [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Build of instance d66e557f-fbd5-4369-bdcb-14f25c3bbc7b was re-scheduled: Binding failed for port 3974d385-edd8-4d22-a7c2-1f79a1c724ec, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 625.671454] env[62965]: DEBUG nova.compute.manager [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 625.671681] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] Acquiring lock "refresh_cache-d66e557f-fbd5-4369-bdcb-14f25c3bbc7b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.671820] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] Acquired lock "refresh_cache-d66e557f-fbd5-4369-bdcb-14f25c3bbc7b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.671971] env[62965]: DEBUG nova.network.neutron [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 625.672954] env[62965]: DEBUG oslo_concurrency.lockutils [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.887s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.695623] env[62965]: DEBUG nova.compute.manager [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 625.731671] env[62965]: DEBUG nova.virt.hardware [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 625.731914] env[62965]: DEBUG nova.virt.hardware [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 625.732077] env[62965]: DEBUG nova.virt.hardware [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 625.732257] env[62965]: DEBUG nova.virt.hardware [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 625.732396] env[62965]: DEBUG nova.virt.hardware [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 625.732535] env[62965]: DEBUG nova.virt.hardware [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 625.732743] env[62965]: DEBUG nova.virt.hardware [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 625.733099] env[62965]: DEBUG nova.virt.hardware [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 625.733277] env[62965]: DEBUG nova.virt.hardware [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 625.733600] env[62965]: DEBUG nova.virt.hardware [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 625.733811] env[62965]: DEBUG nova.virt.hardware [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 625.735365] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fbcdf1a-d552-45db-bb69-1611155be192 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.746161] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9099e0c-d7de-4309-883d-09db48c8ec99 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.762612] env[62965]: DEBUG nova.compute.manager [req-c3e8d919-aa6c-4f83-821e-e41471927899 req-c7c66bd9-92f3-4f67-ad9e-73fbde60bcd2 service nova] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Received event network-changed-b6c21631-dcbf-4d90-ad9b-fa7ddd03e087 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 625.762800] env[62965]: DEBUG nova.compute.manager [req-c3e8d919-aa6c-4f83-821e-e41471927899 req-c7c66bd9-92f3-4f67-ad9e-73fbde60bcd2 service nova] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Refreshing instance network info cache due to event network-changed-b6c21631-dcbf-4d90-ad9b-fa7ddd03e087. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 625.763009] env[62965]: DEBUG oslo_concurrency.lockutils [req-c3e8d919-aa6c-4f83-821e-e41471927899 req-c7c66bd9-92f3-4f67-ad9e-73fbde60bcd2 service nova] Acquiring lock "refresh_cache-09b62788-3082-4c99-9dfc-0843c4988feb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.763159] env[62965]: DEBUG oslo_concurrency.lockutils [req-c3e8d919-aa6c-4f83-821e-e41471927899 req-c7c66bd9-92f3-4f67-ad9e-73fbde60bcd2 service nova] Acquired lock "refresh_cache-09b62788-3082-4c99-9dfc-0843c4988feb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.763315] env[62965]: DEBUG nova.network.neutron [req-c3e8d919-aa6c-4f83-821e-e41471927899 req-c7c66bd9-92f3-4f67-ad9e-73fbde60bcd2 service nova] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Refreshing network info cache for port b6c21631-dcbf-4d90-ad9b-fa7ddd03e087 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 625.921233] env[62965]: DEBUG oslo_concurrency.lockutils [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.005413] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Acquiring lock "f9a2b957-a312-4c9b-bb0d-0b1dad6416cd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.005766] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Lock "f9a2b957-a312-4c9b-bb0d-0b1dad6416cd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 626.066105] env[62965]: ERROR nova.compute.manager [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b6c21631-dcbf-4d90-ad9b-fa7ddd03e087, please check neutron logs for more information. [ 626.066105] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 626.066105] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 626.066105] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 626.066105] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 626.066105] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 626.066105] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 626.066105] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 626.066105] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.066105] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 626.066105] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.066105] env[62965]: ERROR nova.compute.manager raise self.value [ 626.066105] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 626.066105] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 626.066105] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.066105] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 626.066464] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.066464] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 626.066464] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b6c21631-dcbf-4d90-ad9b-fa7ddd03e087, please check neutron logs for more information. [ 626.066464] env[62965]: ERROR nova.compute.manager [ 626.066464] env[62965]: Traceback (most recent call last): [ 626.066464] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 626.066464] env[62965]: listener.cb(fileno) [ 626.066464] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 626.066464] env[62965]: result = function(*args, **kwargs) [ 626.066464] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 626.066464] env[62965]: return func(*args, **kwargs) [ 626.066464] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 626.066464] env[62965]: raise e [ 626.066464] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 626.066464] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 626.066464] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 626.066464] env[62965]: created_port_ids = self._update_ports_for_instance( [ 626.066464] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 626.066464] env[62965]: with excutils.save_and_reraise_exception(): [ 626.066464] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.066464] env[62965]: self.force_reraise() [ 626.066464] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.066464] env[62965]: raise self.value [ 626.066464] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 626.066464] env[62965]: updated_port = self._update_port( [ 626.066464] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.066464] env[62965]: _ensure_no_port_binding_failure(port) [ 626.066464] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.066464] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 626.067076] env[62965]: nova.exception.PortBindingFailed: Binding failed for port b6c21631-dcbf-4d90-ad9b-fa7ddd03e087, please check neutron logs for more information. [ 626.067076] env[62965]: Removing descriptor: 17 [ 626.067076] env[62965]: ERROR nova.compute.manager [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b6c21631-dcbf-4d90-ad9b-fa7ddd03e087, please check neutron logs for more information. [ 626.067076] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Traceback (most recent call last): [ 626.067076] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 626.067076] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] yield resources [ 626.067076] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 626.067076] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] self.driver.spawn(context, instance, image_meta, [ 626.067076] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 626.067076] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 626.067076] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 626.067076] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] vm_ref = self.build_virtual_machine(instance, [ 626.067370] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 626.067370] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] vif_infos = vmwarevif.get_vif_info(self._session, [ 626.067370] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 626.067370] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] for vif in network_info: [ 626.067370] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 626.067370] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] return self._sync_wrapper(fn, *args, **kwargs) [ 626.067370] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 626.067370] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] self.wait() [ 626.067370] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 626.067370] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] self[:] = self._gt.wait() [ 626.067370] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 626.067370] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] return self._exit_event.wait() [ 626.067370] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 626.067646] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] result = hub.switch() [ 626.067646] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 626.067646] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] return self.greenlet.switch() [ 626.067646] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 626.067646] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] result = function(*args, **kwargs) [ 626.067646] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 626.067646] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] return func(*args, **kwargs) [ 626.067646] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 626.067646] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] raise e [ 626.067646] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 626.067646] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] nwinfo = self.network_api.allocate_for_instance( [ 626.067646] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 626.067646] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] created_port_ids = self._update_ports_for_instance( [ 626.067981] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 626.067981] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] with excutils.save_and_reraise_exception(): [ 626.067981] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.067981] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] self.force_reraise() [ 626.067981] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.067981] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] raise self.value [ 626.067981] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 626.067981] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] updated_port = self._update_port( [ 626.067981] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.067981] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] _ensure_no_port_binding_failure(port) [ 626.067981] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.067981] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] raise exception.PortBindingFailed(port_id=port['id']) [ 626.068246] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] nova.exception.PortBindingFailed: Binding failed for port b6c21631-dcbf-4d90-ad9b-fa7ddd03e087, please check neutron logs for more information. [ 626.068246] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] [ 626.068246] env[62965]: INFO nova.compute.manager [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Terminating instance [ 626.199020] env[62965]: DEBUG nova.network.neutron [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 626.254319] env[62965]: DEBUG nova.network.neutron [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.282485] env[62965]: DEBUG nova.network.neutron [req-c3e8d919-aa6c-4f83-821e-e41471927899 req-c7c66bd9-92f3-4f67-ad9e-73fbde60bcd2 service nova] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 626.349900] env[62965]: DEBUG nova.network.neutron [req-c3e8d919-aa6c-4f83-821e-e41471927899 req-c7c66bd9-92f3-4f67-ad9e-73fbde60bcd2 service nova] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.524802] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddbb9569-fa32-4c33-8ac5-3032e1e144c2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.533063] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e7ef35c-bd8d-43ab-9778-a00164f57df9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.562966] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88007ab9-c0df-40d4-9aed-c61050775fa2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.570513] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fba6a0fd-7923-4822-bd0f-90360c929955 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.574484] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Acquiring lock "refresh_cache-09b62788-3082-4c99-9dfc-0843c4988feb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.584061] env[62965]: DEBUG nova.compute.provider_tree [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 626.757136] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] Releasing lock "refresh_cache-d66e557f-fbd5-4369-bdcb-14f25c3bbc7b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.757466] env[62965]: DEBUG nova.compute.manager [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 626.757632] env[62965]: DEBUG nova.compute.manager [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 626.758088] env[62965]: DEBUG nova.network.neutron [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 626.777292] env[62965]: DEBUG nova.network.neutron [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 626.852880] env[62965]: DEBUG oslo_concurrency.lockutils [req-c3e8d919-aa6c-4f83-821e-e41471927899 req-c7c66bd9-92f3-4f67-ad9e-73fbde60bcd2 service nova] Releasing lock "refresh_cache-09b62788-3082-4c99-9dfc-0843c4988feb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.853312] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Acquired lock "refresh_cache-09b62788-3082-4c99-9dfc-0843c4988feb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.853495] env[62965]: DEBUG nova.network.neutron [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 627.087583] env[62965]: DEBUG nova.scheduler.client.report [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 627.280026] env[62965]: DEBUG nova.network.neutron [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.369878] env[62965]: DEBUG nova.network.neutron [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 627.451171] env[62965]: DEBUG nova.network.neutron [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.593300] env[62965]: DEBUG oslo_concurrency.lockutils [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.920s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.594093] env[62965]: ERROR nova.compute.manager [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 68e40c1c-68d7-45b1-8497-f0c9e9619416, please check neutron logs for more information. [ 627.594093] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Traceback (most recent call last): [ 627.594093] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 627.594093] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] self.driver.spawn(context, instance, image_meta, [ 627.594093] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 627.594093] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 627.594093] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 627.594093] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] vm_ref = self.build_virtual_machine(instance, [ 627.594093] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 627.594093] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] vif_infos = vmwarevif.get_vif_info(self._session, [ 627.594093] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 627.594424] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] for vif in network_info: [ 627.594424] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 627.594424] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] return self._sync_wrapper(fn, *args, **kwargs) [ 627.594424] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 627.594424] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] self.wait() [ 627.594424] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 627.594424] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] self[:] = self._gt.wait() [ 627.594424] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 627.594424] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] return self._exit_event.wait() [ 627.594424] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 627.594424] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] result = hub.switch() [ 627.594424] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 627.594424] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] return self.greenlet.switch() [ 627.594779] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 627.594779] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] result = function(*args, **kwargs) [ 627.594779] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 627.594779] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] return func(*args, **kwargs) [ 627.594779] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 627.594779] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] raise e [ 627.594779] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 627.594779] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] nwinfo = self.network_api.allocate_for_instance( [ 627.594779] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 627.594779] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] created_port_ids = self._update_ports_for_instance( [ 627.594779] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 627.594779] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] with excutils.save_and_reraise_exception(): [ 627.594779] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.595131] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] self.force_reraise() [ 627.595131] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.595131] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] raise self.value [ 627.595131] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 627.595131] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] updated_port = self._update_port( [ 627.595131] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.595131] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] _ensure_no_port_binding_failure(port) [ 627.595131] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.595131] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] raise exception.PortBindingFailed(port_id=port['id']) [ 627.595131] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] nova.exception.PortBindingFailed: Binding failed for port 68e40c1c-68d7-45b1-8497-f0c9e9619416, please check neutron logs for more information. [ 627.595131] env[62965]: ERROR nova.compute.manager [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] [ 627.595409] env[62965]: DEBUG nova.compute.utils [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Binding failed for port 68e40c1c-68d7-45b1-8497-f0c9e9619416, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 627.596092] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.595s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.597631] env[62965]: INFO nova.compute.claims [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 627.600290] env[62965]: DEBUG nova.compute.manager [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Build of instance 2931c8b9-ed5f-4994-a4d5-50e1369d67e9 was re-scheduled: Binding failed for port 68e40c1c-68d7-45b1-8497-f0c9e9619416, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 627.600717] env[62965]: DEBUG nova.compute.manager [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 627.600940] env[62965]: DEBUG oslo_concurrency.lockutils [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquiring lock "refresh_cache-2931c8b9-ed5f-4994-a4d5-50e1369d67e9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.601095] env[62965]: DEBUG oslo_concurrency.lockutils [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquired lock "refresh_cache-2931c8b9-ed5f-4994-a4d5-50e1369d67e9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.601249] env[62965]: DEBUG nova.network.neutron [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 627.782371] env[62965]: INFO nova.compute.manager [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] [instance: d66e557f-fbd5-4369-bdcb-14f25c3bbc7b] Took 1.02 seconds to deallocate network for instance. [ 627.923636] env[62965]: DEBUG nova.compute.manager [req-ca538351-579d-444e-b469-0bcbe726a7c7 req-b856d922-3c9e-49ef-83f3-e065a5f54294 service nova] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Received event network-vif-deleted-b6c21631-dcbf-4d90-ad9b-fa7ddd03e087 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 627.954027] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Releasing lock "refresh_cache-09b62788-3082-4c99-9dfc-0843c4988feb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.954472] env[62965]: DEBUG nova.compute.manager [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 627.954660] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 627.954948] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c5fe61f0-48a6-41a9-85e6-ede7648b0f18 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.964240] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0e0e2f5-0a81-4dbd-bc5b-85c8e94787c1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.985756] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 09b62788-3082-4c99-9dfc-0843c4988feb could not be found. [ 627.985969] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 627.986163] env[62965]: INFO nova.compute.manager [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Took 0.03 seconds to destroy the instance on the hypervisor. [ 627.986403] env[62965]: DEBUG oslo.service.loopingcall [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 627.986657] env[62965]: DEBUG nova.compute.manager [-] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 627.986707] env[62965]: DEBUG nova.network.neutron [-] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 628.008783] env[62965]: DEBUG nova.network.neutron [-] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.120219] env[62965]: DEBUG nova.network.neutron [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.231346] env[62965]: DEBUG nova.network.neutron [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.511580] env[62965]: DEBUG nova.network.neutron [-] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.736426] env[62965]: DEBUG oslo_concurrency.lockutils [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Releasing lock "refresh_cache-2931c8b9-ed5f-4994-a4d5-50e1369d67e9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.736647] env[62965]: DEBUG nova.compute.manager [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 628.736825] env[62965]: DEBUG nova.compute.manager [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 628.736991] env[62965]: DEBUG nova.network.neutron [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 628.755133] env[62965]: DEBUG nova.network.neutron [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.812232] env[62965]: INFO nova.scheduler.client.report [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] Deleted allocations for instance d66e557f-fbd5-4369-bdcb-14f25c3bbc7b [ 628.973214] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a6e9a45-8144-40ae-b44c-1f723a4192b1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.980784] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ecf1399-0e51-46da-b3fd-c827b9200787 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.012036] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41daefaa-1bf2-4bd8-a0c6-d67d5ad3d20c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.014954] env[62965]: INFO nova.compute.manager [-] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Took 1.03 seconds to deallocate network for instance. [ 629.017239] env[62965]: DEBUG nova.compute.claims [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 629.017393] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.021681] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d20b9e9c-a396-49d7-8fed-cbf3576ebc05 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.035105] env[62965]: DEBUG nova.compute.provider_tree [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 629.258449] env[62965]: DEBUG nova.network.neutron [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.321585] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2ebde45c-aee8-42fe-aa22-8540160c0199 tempest-TenantUsagesTestJSON-1031246219 tempest-TenantUsagesTestJSON-1031246219-project-member] Lock "d66e557f-fbd5-4369-bdcb-14f25c3bbc7b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.502s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.538771] env[62965]: DEBUG nova.scheduler.client.report [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 629.760999] env[62965]: INFO nova.compute.manager [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 2931c8b9-ed5f-4994-a4d5-50e1369d67e9] Took 1.02 seconds to deallocate network for instance. [ 629.824800] env[62965]: DEBUG nova.compute.manager [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 630.046321] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.450s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.046876] env[62965]: DEBUG nova.compute.manager [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 630.051863] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.716s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.352241] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.556999] env[62965]: DEBUG nova.compute.utils [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 630.561427] env[62965]: DEBUG nova.compute.manager [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 630.561592] env[62965]: DEBUG nova.network.neutron [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 630.628112] env[62965]: DEBUG nova.policy [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f16e1775265f4ec3adbcd66d4df484a3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '503e420565454b169ca53f0779a040af', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 630.795990] env[62965]: INFO nova.scheduler.client.report [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Deleted allocations for instance 2931c8b9-ed5f-4994-a4d5-50e1369d67e9 [ 630.936269] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af235459-a988-4d41-b65b-baa7eb22385a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.944235] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7cec6ac-bab9-47cd-8bda-ff88c3dabb51 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.975659] env[62965]: DEBUG nova.network.neutron [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Successfully created port: c768e9b3-53f1-42b5-b16e-edd046109356 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 630.978496] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fc0a568-76ad-46e3-af6c-6cfd439afda0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.985769] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba507769-6070-42f6-a531-9ce0cabe0c10 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.000664] env[62965]: DEBUG nova.compute.provider_tree [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 631.062147] env[62965]: DEBUG nova.compute.manager [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 631.308381] env[62965]: DEBUG oslo_concurrency.lockutils [None req-12e7fde8-a680-4676-8ed6-60f416f407a7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "2931c8b9-ed5f-4994-a4d5-50e1369d67e9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.031s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.505019] env[62965]: DEBUG nova.scheduler.client.report [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 631.813594] env[62965]: DEBUG nova.compute.manager [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 632.010875] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.959s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 632.011548] env[62965]: ERROR nova.compute.manager [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0e5fdb64-ae19-4373-87ba-620ccd0339f2, please check neutron logs for more information. [ 632.011548] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Traceback (most recent call last): [ 632.011548] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 632.011548] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] self.driver.spawn(context, instance, image_meta, [ 632.011548] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 632.011548] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] self._vmops.spawn(context, instance, image_meta, injected_files, [ 632.011548] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 632.011548] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] vm_ref = self.build_virtual_machine(instance, [ 632.011548] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 632.011548] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] vif_infos = vmwarevif.get_vif_info(self._session, [ 632.011548] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 632.011901] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] for vif in network_info: [ 632.011901] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 632.011901] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] return self._sync_wrapper(fn, *args, **kwargs) [ 632.011901] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 632.011901] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] self.wait() [ 632.011901] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 632.011901] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] self[:] = self._gt.wait() [ 632.011901] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 632.011901] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] return self._exit_event.wait() [ 632.011901] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 632.011901] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] result = hub.switch() [ 632.011901] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 632.011901] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] return self.greenlet.switch() [ 632.012329] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.012329] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] result = function(*args, **kwargs) [ 632.012329] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 632.012329] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] return func(*args, **kwargs) [ 632.012329] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 632.012329] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] raise e [ 632.012329] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 632.012329] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] nwinfo = self.network_api.allocate_for_instance( [ 632.012329] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 632.012329] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] created_port_ids = self._update_ports_for_instance( [ 632.012329] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 632.012329] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] with excutils.save_and_reraise_exception(): [ 632.012329] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.012718] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] self.force_reraise() [ 632.012718] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.012718] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] raise self.value [ 632.012718] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 632.012718] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] updated_port = self._update_port( [ 632.012718] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.012718] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] _ensure_no_port_binding_failure(port) [ 632.012718] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.012718] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] raise exception.PortBindingFailed(port_id=port['id']) [ 632.012718] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] nova.exception.PortBindingFailed: Binding failed for port 0e5fdb64-ae19-4373-87ba-620ccd0339f2, please check neutron logs for more information. [ 632.012718] env[62965]: ERROR nova.compute.manager [instance: f619dd08-0cdf-486f-8f13-707b254a8593] [ 632.013030] env[62965]: DEBUG nova.compute.utils [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Binding failed for port 0e5fdb64-ae19-4373-87ba-620ccd0339f2, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 632.016855] env[62965]: DEBUG oslo_concurrency.lockutils [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.600s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 632.019890] env[62965]: DEBUG nova.compute.manager [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Build of instance f619dd08-0cdf-486f-8f13-707b254a8593 was re-scheduled: Binding failed for port 0e5fdb64-ae19-4373-87ba-620ccd0339f2, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 632.020348] env[62965]: DEBUG nova.compute.manager [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 632.020565] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Acquiring lock "refresh_cache-f619dd08-0cdf-486f-8f13-707b254a8593" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 632.020707] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Acquired lock "refresh_cache-f619dd08-0cdf-486f-8f13-707b254a8593" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.020893] env[62965]: DEBUG nova.network.neutron [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 632.071729] env[62965]: DEBUG nova.compute.manager [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 632.096835] env[62965]: DEBUG nova.virt.hardware [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 632.096835] env[62965]: DEBUG nova.virt.hardware [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 632.097027] env[62965]: DEBUG nova.virt.hardware [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 632.099023] env[62965]: DEBUG nova.virt.hardware [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 632.099023] env[62965]: DEBUG nova.virt.hardware [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 632.099023] env[62965]: DEBUG nova.virt.hardware [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 632.099023] env[62965]: DEBUG nova.virt.hardware [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 632.099023] env[62965]: DEBUG nova.virt.hardware [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 632.099389] env[62965]: DEBUG nova.virt.hardware [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 632.099389] env[62965]: DEBUG nova.virt.hardware [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 632.099389] env[62965]: DEBUG nova.virt.hardware [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 632.099389] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c20aee4f-3266-4f1b-897f-12a8b0cb8f4e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.107144] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad81be2c-f0bf-4649-99e6-864ea2fbd0c1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.150437] env[62965]: DEBUG nova.compute.manager [req-934b2c9e-8203-45ae-8a35-bc72b61592fa req-dfe2147e-4e57-476a-a289-5e5823df092e service nova] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Received event network-changed-c768e9b3-53f1-42b5-b16e-edd046109356 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 632.150437] env[62965]: DEBUG nova.compute.manager [req-934b2c9e-8203-45ae-8a35-bc72b61592fa req-dfe2147e-4e57-476a-a289-5e5823df092e service nova] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Refreshing instance network info cache due to event network-changed-c768e9b3-53f1-42b5-b16e-edd046109356. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 632.150437] env[62965]: DEBUG oslo_concurrency.lockutils [req-934b2c9e-8203-45ae-8a35-bc72b61592fa req-dfe2147e-4e57-476a-a289-5e5823df092e service nova] Acquiring lock "refresh_cache-63c54ed4-4525-42f6-a99f-033f7b9838a7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 632.150437] env[62965]: DEBUG oslo_concurrency.lockutils [req-934b2c9e-8203-45ae-8a35-bc72b61592fa req-dfe2147e-4e57-476a-a289-5e5823df092e service nova] Acquired lock "refresh_cache-63c54ed4-4525-42f6-a99f-033f7b9838a7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.150437] env[62965]: DEBUG nova.network.neutron [req-934b2c9e-8203-45ae-8a35-bc72b61592fa req-dfe2147e-4e57-476a-a289-5e5823df092e service nova] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Refreshing network info cache for port c768e9b3-53f1-42b5-b16e-edd046109356 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 632.345745] env[62965]: DEBUG oslo_concurrency.lockutils [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.433631] env[62965]: ERROR nova.compute.manager [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c768e9b3-53f1-42b5-b16e-edd046109356, please check neutron logs for more information. [ 632.433631] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 632.433631] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 632.433631] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 632.433631] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 632.433631] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 632.433631] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 632.433631] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 632.433631] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.433631] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 632.433631] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.433631] env[62965]: ERROR nova.compute.manager raise self.value [ 632.433631] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 632.433631] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 632.433631] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.433631] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 632.434097] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.434097] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 632.434097] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c768e9b3-53f1-42b5-b16e-edd046109356, please check neutron logs for more information. [ 632.434097] env[62965]: ERROR nova.compute.manager [ 632.434097] env[62965]: Traceback (most recent call last): [ 632.434097] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 632.434097] env[62965]: listener.cb(fileno) [ 632.434097] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.434097] env[62965]: result = function(*args, **kwargs) [ 632.434097] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 632.434097] env[62965]: return func(*args, **kwargs) [ 632.434097] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 632.434097] env[62965]: raise e [ 632.434097] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 632.434097] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 632.434097] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 632.434097] env[62965]: created_port_ids = self._update_ports_for_instance( [ 632.434097] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 632.434097] env[62965]: with excutils.save_and_reraise_exception(): [ 632.434097] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.434097] env[62965]: self.force_reraise() [ 632.434097] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.434097] env[62965]: raise self.value [ 632.434097] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 632.434097] env[62965]: updated_port = self._update_port( [ 632.434097] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.434097] env[62965]: _ensure_no_port_binding_failure(port) [ 632.434097] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.434097] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 632.434946] env[62965]: nova.exception.PortBindingFailed: Binding failed for port c768e9b3-53f1-42b5-b16e-edd046109356, please check neutron logs for more information. [ 632.434946] env[62965]: Removing descriptor: 17 [ 632.434946] env[62965]: ERROR nova.compute.manager [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c768e9b3-53f1-42b5-b16e-edd046109356, please check neutron logs for more information. [ 632.434946] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Traceback (most recent call last): [ 632.434946] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 632.434946] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] yield resources [ 632.434946] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 632.434946] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] self.driver.spawn(context, instance, image_meta, [ 632.434946] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 632.434946] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 632.434946] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 632.434946] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] vm_ref = self.build_virtual_machine(instance, [ 632.435279] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 632.435279] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] vif_infos = vmwarevif.get_vif_info(self._session, [ 632.435279] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 632.435279] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] for vif in network_info: [ 632.435279] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 632.435279] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] return self._sync_wrapper(fn, *args, **kwargs) [ 632.435279] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 632.435279] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] self.wait() [ 632.435279] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 632.435279] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] self[:] = self._gt.wait() [ 632.435279] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 632.435279] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] return self._exit_event.wait() [ 632.435279] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 632.435662] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] result = hub.switch() [ 632.435662] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 632.435662] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] return self.greenlet.switch() [ 632.435662] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.435662] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] result = function(*args, **kwargs) [ 632.435662] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 632.435662] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] return func(*args, **kwargs) [ 632.435662] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 632.435662] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] raise e [ 632.435662] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 632.435662] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] nwinfo = self.network_api.allocate_for_instance( [ 632.435662] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 632.435662] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] created_port_ids = self._update_ports_for_instance( [ 632.436096] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 632.436096] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] with excutils.save_and_reraise_exception(): [ 632.436096] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.436096] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] self.force_reraise() [ 632.436096] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.436096] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] raise self.value [ 632.436096] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 632.436096] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] updated_port = self._update_port( [ 632.436096] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.436096] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] _ensure_no_port_binding_failure(port) [ 632.436096] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.436096] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] raise exception.PortBindingFailed(port_id=port['id']) [ 632.436486] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] nova.exception.PortBindingFailed: Binding failed for port c768e9b3-53f1-42b5-b16e-edd046109356, please check neutron logs for more information. [ 632.436486] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] [ 632.436486] env[62965]: INFO nova.compute.manager [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Terminating instance [ 632.541659] env[62965]: DEBUG nova.network.neutron [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.659276] env[62965]: DEBUG nova.network.neutron [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.671345] env[62965]: DEBUG nova.network.neutron [req-934b2c9e-8203-45ae-8a35-bc72b61592fa req-dfe2147e-4e57-476a-a289-5e5823df092e service nova] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.744416] env[62965]: DEBUG nova.network.neutron [req-934b2c9e-8203-45ae-8a35-bc72b61592fa req-dfe2147e-4e57-476a-a289-5e5823df092e service nova] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.937386] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquiring lock "3a994542-8c8e-4b3c-87bc-2c40a0fc95b5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.937671] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "3a994542-8c8e-4b3c-87bc-2c40a0fc95b5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 632.939396] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] Acquiring lock "refresh_cache-63c54ed4-4525-42f6-a99f-033f7b9838a7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 632.955142] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73be6577-4979-40cc-bd34-77bf2374f015 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.962574] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c588952-d0f5-4934-846f-ecae018828e4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.992691] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58bc2493-de33-4170-a5cf-e1208a5aecb3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.999667] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4be53b1f-5ea8-4f50-911a-b5f226606b71 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.012449] env[62965]: DEBUG nova.compute.provider_tree [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 633.163410] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Releasing lock "refresh_cache-f619dd08-0cdf-486f-8f13-707b254a8593" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 633.163661] env[62965]: DEBUG nova.compute.manager [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 633.163841] env[62965]: DEBUG nova.compute.manager [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 633.164014] env[62965]: DEBUG nova.network.neutron [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 633.179543] env[62965]: DEBUG nova.network.neutron [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.247429] env[62965]: DEBUG oslo_concurrency.lockutils [req-934b2c9e-8203-45ae-8a35-bc72b61592fa req-dfe2147e-4e57-476a-a289-5e5823df092e service nova] Releasing lock "refresh_cache-63c54ed4-4525-42f6-a99f-033f7b9838a7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 633.247848] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] Acquired lock "refresh_cache-63c54ed4-4525-42f6-a99f-033f7b9838a7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.248042] env[62965]: DEBUG nova.network.neutron [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 633.515091] env[62965]: DEBUG nova.scheduler.client.report [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 633.682187] env[62965]: DEBUG nova.network.neutron [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.765642] env[62965]: DEBUG nova.network.neutron [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.837809] env[62965]: DEBUG nova.network.neutron [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.020260] env[62965]: DEBUG oslo_concurrency.lockutils [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.004s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 634.020555] env[62965]: ERROR nova.compute.manager [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ce89fee2-26e8-4f78-9de6-148d957ccb38, please check neutron logs for more information. [ 634.020555] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Traceback (most recent call last): [ 634.020555] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 634.020555] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] self.driver.spawn(context, instance, image_meta, [ 634.020555] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 634.020555] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] self._vmops.spawn(context, instance, image_meta, injected_files, [ 634.020555] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 634.020555] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] vm_ref = self.build_virtual_machine(instance, [ 634.020555] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 634.020555] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] vif_infos = vmwarevif.get_vif_info(self._session, [ 634.020555] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 634.020875] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] for vif in network_info: [ 634.020875] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 634.020875] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] return self._sync_wrapper(fn, *args, **kwargs) [ 634.020875] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 634.020875] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] self.wait() [ 634.020875] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 634.020875] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] self[:] = self._gt.wait() [ 634.020875] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 634.020875] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] return self._exit_event.wait() [ 634.020875] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 634.020875] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] result = hub.switch() [ 634.020875] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 634.020875] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] return self.greenlet.switch() [ 634.021217] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 634.021217] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] result = function(*args, **kwargs) [ 634.021217] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 634.021217] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] return func(*args, **kwargs) [ 634.021217] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 634.021217] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] raise e [ 634.021217] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 634.021217] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] nwinfo = self.network_api.allocate_for_instance( [ 634.021217] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 634.021217] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] created_port_ids = self._update_ports_for_instance( [ 634.021217] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 634.021217] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] with excutils.save_and_reraise_exception(): [ 634.021217] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 634.021559] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] self.force_reraise() [ 634.021559] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 634.021559] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] raise self.value [ 634.021559] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 634.021559] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] updated_port = self._update_port( [ 634.021559] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 634.021559] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] _ensure_no_port_binding_failure(port) [ 634.021559] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 634.021559] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] raise exception.PortBindingFailed(port_id=port['id']) [ 634.021559] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] nova.exception.PortBindingFailed: Binding failed for port ce89fee2-26e8-4f78-9de6-148d957ccb38, please check neutron logs for more information. [ 634.021559] env[62965]: ERROR nova.compute.manager [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] [ 634.021860] env[62965]: DEBUG nova.compute.utils [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Binding failed for port ce89fee2-26e8-4f78-9de6-148d957ccb38, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 634.022789] env[62965]: DEBUG oslo_concurrency.lockutils [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.858s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.025651] env[62965]: DEBUG nova.compute.manager [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Build of instance 61702d54-59d2-4d13-9eed-59b052b7fe74 was re-scheduled: Binding failed for port ce89fee2-26e8-4f78-9de6-148d957ccb38, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 634.026208] env[62965]: DEBUG nova.compute.manager [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 634.026438] env[62965]: DEBUG oslo_concurrency.lockutils [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] Acquiring lock "refresh_cache-61702d54-59d2-4d13-9eed-59b052b7fe74" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 634.026583] env[62965]: DEBUG oslo_concurrency.lockutils [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] Acquired lock "refresh_cache-61702d54-59d2-4d13-9eed-59b052b7fe74" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.026739] env[62965]: DEBUG nova.network.neutron [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 634.173800] env[62965]: DEBUG nova.compute.manager [req-6bf4e149-1581-4f34-9216-3f490efc62d4 req-66801fdb-c9f2-4f3f-a4d7-4da99e4045cb service nova] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Received event network-vif-deleted-c768e9b3-53f1-42b5-b16e-edd046109356 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 634.185274] env[62965]: INFO nova.compute.manager [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: f619dd08-0cdf-486f-8f13-707b254a8593] Took 1.02 seconds to deallocate network for instance. [ 634.340271] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] Releasing lock "refresh_cache-63c54ed4-4525-42f6-a99f-033f7b9838a7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.340697] env[62965]: DEBUG nova.compute.manager [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 634.340894] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 634.341220] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f0170157-8608-45d6-af75-f0c7e06c0022 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.356233] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83875d9c-9340-4f71-a9bf-9b001ed573c8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.379902] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 63c54ed4-4525-42f6-a99f-033f7b9838a7 could not be found. [ 634.380126] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 634.380300] env[62965]: INFO nova.compute.manager [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Took 0.04 seconds to destroy the instance on the hypervisor. [ 634.380533] env[62965]: DEBUG oslo.service.loopingcall [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 634.380750] env[62965]: DEBUG nova.compute.manager [-] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 634.380845] env[62965]: DEBUG nova.network.neutron [-] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 634.395136] env[62965]: DEBUG nova.network.neutron [-] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 634.547082] env[62965]: DEBUG nova.network.neutron [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 634.636810] env[62965]: DEBUG nova.network.neutron [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.897960] env[62965]: DEBUG nova.network.neutron [-] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.942551] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc25f78d-69f3-4eb1-ba3e-7450eff23372 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.950911] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0535e367-a747-485a-8983-a41e693aac84 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.551399] env[62965]: DEBUG oslo_concurrency.lockutils [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] Releasing lock "refresh_cache-61702d54-59d2-4d13-9eed-59b052b7fe74" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 635.551667] env[62965]: DEBUG nova.compute.manager [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 635.551795] env[62965]: DEBUG nova.compute.manager [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 635.551955] env[62965]: DEBUG nova.network.neutron [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 635.556057] env[62965]: INFO nova.compute.manager [-] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Took 1.18 seconds to deallocate network for instance. [ 635.558467] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e46f41b-30f3-42b4-bd81-9bb03b7630ed {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.561019] env[62965]: DEBUG nova.compute.claims [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 635.561201] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.567434] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b20610b9-4be3-45e3-939b-8a7c0d1a8fba {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.581088] env[62965]: DEBUG nova.compute.provider_tree [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 635.583035] env[62965]: DEBUG nova.network.neutron [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 635.584740] env[62965]: INFO nova.scheduler.client.report [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Deleted allocations for instance f619dd08-0cdf-486f-8f13-707b254a8593 [ 636.095607] env[62965]: DEBUG nova.scheduler.client.report [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 636.099157] env[62965]: DEBUG nova.network.neutron [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.102793] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1bf5721a-1e86-46c1-ab12-3efc91fd666f tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Lock "f619dd08-0cdf-486f-8f13-707b254a8593" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.788s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 636.601732] env[62965]: DEBUG oslo_concurrency.lockutils [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.579s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 636.602382] env[62965]: ERROR nova.compute.manager [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port adf3e41d-5e61-406e-b35a-fb511e5bfcaa, please check neutron logs for more information. [ 636.602382] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Traceback (most recent call last): [ 636.602382] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 636.602382] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] self.driver.spawn(context, instance, image_meta, [ 636.602382] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 636.602382] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] self._vmops.spawn(context, instance, image_meta, injected_files, [ 636.602382] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 636.602382] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] vm_ref = self.build_virtual_machine(instance, [ 636.602382] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 636.602382] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] vif_infos = vmwarevif.get_vif_info(self._session, [ 636.602382] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 636.602712] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] for vif in network_info: [ 636.602712] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 636.602712] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] return self._sync_wrapper(fn, *args, **kwargs) [ 636.602712] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 636.602712] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] self.wait() [ 636.602712] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 636.602712] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] self[:] = self._gt.wait() [ 636.602712] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 636.602712] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] return self._exit_event.wait() [ 636.602712] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 636.602712] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] result = hub.switch() [ 636.602712] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 636.602712] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] return self.greenlet.switch() [ 636.603075] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 636.603075] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] result = function(*args, **kwargs) [ 636.603075] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 636.603075] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] return func(*args, **kwargs) [ 636.603075] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 636.603075] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] raise e [ 636.603075] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 636.603075] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] nwinfo = self.network_api.allocate_for_instance( [ 636.603075] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 636.603075] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] created_port_ids = self._update_ports_for_instance( [ 636.603075] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 636.603075] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] with excutils.save_and_reraise_exception(): [ 636.603075] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 636.603435] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] self.force_reraise() [ 636.603435] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 636.603435] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] raise self.value [ 636.603435] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 636.603435] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] updated_port = self._update_port( [ 636.603435] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 636.603435] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] _ensure_no_port_binding_failure(port) [ 636.603435] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 636.603435] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] raise exception.PortBindingFailed(port_id=port['id']) [ 636.603435] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] nova.exception.PortBindingFailed: Binding failed for port adf3e41d-5e61-406e-b35a-fb511e5bfcaa, please check neutron logs for more information. [ 636.603435] env[62965]: ERROR nova.compute.manager [instance: 59187ca2-0482-4603-8689-542d2bf5d023] [ 636.603731] env[62965]: DEBUG nova.compute.utils [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Binding failed for port adf3e41d-5e61-406e-b35a-fb511e5bfcaa, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 636.604295] env[62965]: DEBUG oslo_concurrency.lockutils [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.937s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 636.607276] env[62965]: INFO nova.compute.manager [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] [instance: 61702d54-59d2-4d13-9eed-59b052b7fe74] Took 1.06 seconds to deallocate network for instance. [ 636.609850] env[62965]: DEBUG nova.compute.manager [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Build of instance 59187ca2-0482-4603-8689-542d2bf5d023 was re-scheduled: Binding failed for port adf3e41d-5e61-406e-b35a-fb511e5bfcaa, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 636.610341] env[62965]: DEBUG nova.compute.manager [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 636.610565] env[62965]: DEBUG oslo_concurrency.lockutils [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] Acquiring lock "refresh_cache-59187ca2-0482-4603-8689-542d2bf5d023" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.610708] env[62965]: DEBUG oslo_concurrency.lockutils [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] Acquired lock "refresh_cache-59187ca2-0482-4603-8689-542d2bf5d023" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.610862] env[62965]: DEBUG nova.network.neutron [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 636.611910] env[62965]: DEBUG nova.compute.manager [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 637.148887] env[62965]: DEBUG nova.network.neutron [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 637.160776] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.238482] env[62965]: DEBUG nova.network.neutron [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.346903] env[62965]: DEBUG oslo_concurrency.lockutils [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Acquiring lock "864ce643-8620-4405-b345-9164c81fec2a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.347347] env[62965]: DEBUG oslo_concurrency.lockutils [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Lock "864ce643-8620-4405-b345-9164c81fec2a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.502995] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7f7c4a5-7d0e-45d6-a5c3-20a38c0847a6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.510856] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfa9d5af-2ede-4316-b394-0e20d937cb68 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.541898] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7041f29e-c679-4d25-8aa7-5d3f00558246 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.549222] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afd97ef2-3c35-4091-81c0-46a0325d4f66 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.562462] env[62965]: DEBUG nova.compute.provider_tree [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 637.590135] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Acquiring lock "d6e90043-ebd3-4daf-acfa-545ce8462942" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.590360] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Lock "d6e90043-ebd3-4daf-acfa-545ce8462942" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.647442] env[62965]: INFO nova.scheduler.client.report [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] Deleted allocations for instance 61702d54-59d2-4d13-9eed-59b052b7fe74 [ 637.745360] env[62965]: DEBUG oslo_concurrency.lockutils [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] Releasing lock "refresh_cache-59187ca2-0482-4603-8689-542d2bf5d023" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 637.745607] env[62965]: DEBUG nova.compute.manager [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 637.745802] env[62965]: DEBUG nova.compute.manager [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 637.745967] env[62965]: DEBUG nova.network.neutron [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 637.761216] env[62965]: DEBUG nova.network.neutron [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 638.066879] env[62965]: DEBUG nova.scheduler.client.report [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 638.155592] env[62965]: DEBUG oslo_concurrency.lockutils [None req-70efc731-9ac1-44f5-bb59-698340dc115c tempest-VolumesAssistedSnapshotsTest-1467083278 tempest-VolumesAssistedSnapshotsTest-1467083278-project-member] Lock "61702d54-59d2-4d13-9eed-59b052b7fe74" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 79.217s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 638.264052] env[62965]: DEBUG nova.network.neutron [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.572040] env[62965]: DEBUG oslo_concurrency.lockutils [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.968s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 638.572607] env[62965]: ERROR nova.compute.manager [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b5e4287e-3c4b-4299-a138-4be56eede9c4, please check neutron logs for more information. [ 638.572607] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Traceback (most recent call last): [ 638.572607] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 638.572607] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] self.driver.spawn(context, instance, image_meta, [ 638.572607] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 638.572607] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 638.572607] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 638.572607] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] vm_ref = self.build_virtual_machine(instance, [ 638.572607] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 638.572607] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] vif_infos = vmwarevif.get_vif_info(self._session, [ 638.572607] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 638.572922] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] for vif in network_info: [ 638.572922] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 638.572922] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] return self._sync_wrapper(fn, *args, **kwargs) [ 638.572922] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 638.572922] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] self.wait() [ 638.572922] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 638.572922] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] self[:] = self._gt.wait() [ 638.572922] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 638.572922] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] return self._exit_event.wait() [ 638.572922] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 638.572922] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] result = hub.switch() [ 638.572922] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 638.572922] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] return self.greenlet.switch() [ 638.573260] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 638.573260] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] result = function(*args, **kwargs) [ 638.573260] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 638.573260] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] return func(*args, **kwargs) [ 638.573260] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 638.573260] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] raise e [ 638.573260] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 638.573260] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] nwinfo = self.network_api.allocate_for_instance( [ 638.573260] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 638.573260] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] created_port_ids = self._update_ports_for_instance( [ 638.573260] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 638.573260] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] with excutils.save_and_reraise_exception(): [ 638.573260] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.573594] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] self.force_reraise() [ 638.573594] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.573594] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] raise self.value [ 638.573594] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 638.573594] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] updated_port = self._update_port( [ 638.573594] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.573594] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] _ensure_no_port_binding_failure(port) [ 638.573594] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.573594] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] raise exception.PortBindingFailed(port_id=port['id']) [ 638.573594] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] nova.exception.PortBindingFailed: Binding failed for port b5e4287e-3c4b-4299-a138-4be56eede9c4, please check neutron logs for more information. [ 638.573594] env[62965]: ERROR nova.compute.manager [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] [ 638.573898] env[62965]: DEBUG nova.compute.utils [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Binding failed for port b5e4287e-3c4b-4299-a138-4be56eede9c4, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 638.574653] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.505s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 638.577905] env[62965]: DEBUG nova.compute.manager [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Build of instance 53519d3e-34d2-410f-ac19-69a015e407b9 was re-scheduled: Binding failed for port b5e4287e-3c4b-4299-a138-4be56eede9c4, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 638.578385] env[62965]: DEBUG nova.compute.manager [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 638.578732] env[62965]: DEBUG oslo_concurrency.lockutils [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] Acquiring lock "refresh_cache-53519d3e-34d2-410f-ac19-69a015e407b9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.578770] env[62965]: DEBUG oslo_concurrency.lockutils [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] Acquired lock "refresh_cache-53519d3e-34d2-410f-ac19-69a015e407b9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.578959] env[62965]: DEBUG nova.network.neutron [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 638.658567] env[62965]: DEBUG nova.compute.manager [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 638.766674] env[62965]: INFO nova.compute.manager [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] [instance: 59187ca2-0482-4603-8689-542d2bf5d023] Took 1.02 seconds to deallocate network for instance. [ 639.101923] env[62965]: DEBUG nova.network.neutron [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.180792] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.183089] env[62965]: DEBUG nova.network.neutron [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.452087] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ecd2a8c-230a-4618-a063-218e437c833c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.459381] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42ed47cd-3576-4663-990f-21c3b4486516 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.489180] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a3d4a2d-602a-4e57-bddb-b53c916e210e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.496134] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-559d4070-fa3e-4ca9-9b25-7c40aa7861d6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.508833] env[62965]: DEBUG nova.compute.provider_tree [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 639.684769] env[62965]: DEBUG oslo_concurrency.lockutils [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] Releasing lock "refresh_cache-53519d3e-34d2-410f-ac19-69a015e407b9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.685054] env[62965]: DEBUG nova.compute.manager [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 639.685054] env[62965]: DEBUG nova.compute.manager [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 639.685207] env[62965]: DEBUG nova.network.neutron [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 639.704712] env[62965]: DEBUG nova.network.neutron [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.798220] env[62965]: INFO nova.scheduler.client.report [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] Deleted allocations for instance 59187ca2-0482-4603-8689-542d2bf5d023 [ 640.011572] env[62965]: DEBUG nova.scheduler.client.report [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 640.184016] env[62965]: DEBUG oslo_concurrency.lockutils [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Acquiring lock "afaec369-6aef-4bdc-b886-eed795644867" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.184016] env[62965]: DEBUG oslo_concurrency.lockutils [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Lock "afaec369-6aef-4bdc-b886-eed795644867" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.208540] env[62965]: DEBUG nova.network.neutron [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.308546] env[62965]: DEBUG oslo_concurrency.lockutils [None req-71506b46-bdd1-404e-8ff3-96f7ab19ab13 tempest-ServersWithSpecificFlavorTestJSON-692094112 tempest-ServersWithSpecificFlavorTestJSON-692094112-project-member] Lock "59187ca2-0482-4603-8689-542d2bf5d023" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 79.971s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 640.516148] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.941s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 640.516768] env[62965]: ERROR nova.compute.manager [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8dc9700a-cf9b-4c75-b4a1-e1c34262d46e, please check neutron logs for more information. [ 640.516768] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Traceback (most recent call last): [ 640.516768] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 640.516768] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] self.driver.spawn(context, instance, image_meta, [ 640.516768] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 640.516768] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 640.516768] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 640.516768] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] vm_ref = self.build_virtual_machine(instance, [ 640.516768] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 640.516768] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] vif_infos = vmwarevif.get_vif_info(self._session, [ 640.516768] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 640.517119] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] for vif in network_info: [ 640.517119] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 640.517119] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] return self._sync_wrapper(fn, *args, **kwargs) [ 640.517119] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 640.517119] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] self.wait() [ 640.517119] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 640.517119] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] self[:] = self._gt.wait() [ 640.517119] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 640.517119] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] return self._exit_event.wait() [ 640.517119] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 640.517119] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] current.throw(*self._exc) [ 640.517119] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 640.517119] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] result = function(*args, **kwargs) [ 640.517482] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 640.517482] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] return func(*args, **kwargs) [ 640.517482] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 640.517482] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] raise e [ 640.517482] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 640.517482] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] nwinfo = self.network_api.allocate_for_instance( [ 640.517482] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 640.517482] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] created_port_ids = self._update_ports_for_instance( [ 640.517482] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 640.517482] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] with excutils.save_and_reraise_exception(): [ 640.517482] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.517482] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] self.force_reraise() [ 640.517482] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.517894] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] raise self.value [ 640.517894] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 640.517894] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] updated_port = self._update_port( [ 640.517894] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.517894] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] _ensure_no_port_binding_failure(port) [ 640.517894] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.517894] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] raise exception.PortBindingFailed(port_id=port['id']) [ 640.517894] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] nova.exception.PortBindingFailed: Binding failed for port 8dc9700a-cf9b-4c75-b4a1-e1c34262d46e, please check neutron logs for more information. [ 640.517894] env[62965]: ERROR nova.compute.manager [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] [ 640.517894] env[62965]: DEBUG nova.compute.utils [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Binding failed for port 8dc9700a-cf9b-4c75-b4a1-e1c34262d46e, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 640.518776] env[62965]: DEBUG oslo_concurrency.lockutils [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.598s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.520569] env[62965]: INFO nova.compute.claims [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 640.523106] env[62965]: DEBUG nova.compute.manager [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Build of instance f594c50e-25b3-4e85-840c-7dfe0994a2c2 was re-scheduled: Binding failed for port 8dc9700a-cf9b-4c75-b4a1-e1c34262d46e, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 640.523537] env[62965]: DEBUG nova.compute.manager [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 640.523757] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] Acquiring lock "refresh_cache-f594c50e-25b3-4e85-840c-7dfe0994a2c2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 640.523899] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] Acquired lock "refresh_cache-f594c50e-25b3-4e85-840c-7dfe0994a2c2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.524111] env[62965]: DEBUG nova.network.neutron [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 640.711510] env[62965]: INFO nova.compute.manager [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] [instance: 53519d3e-34d2-410f-ac19-69a015e407b9] Took 1.03 seconds to deallocate network for instance. [ 640.813085] env[62965]: DEBUG nova.compute.manager [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 641.057531] env[62965]: DEBUG nova.network.neutron [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 641.143680] env[62965]: DEBUG nova.network.neutron [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.337360] env[62965]: DEBUG oslo_concurrency.lockutils [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.372704] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 641.372704] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 641.645998] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] Releasing lock "refresh_cache-f594c50e-25b3-4e85-840c-7dfe0994a2c2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 641.646243] env[62965]: DEBUG nova.compute.manager [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 641.646423] env[62965]: DEBUG nova.compute.manager [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 641.646663] env[62965]: DEBUG nova.network.neutron [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 641.685574] env[62965]: DEBUG nova.network.neutron [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 641.741884] env[62965]: INFO nova.scheduler.client.report [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] Deleted allocations for instance 53519d3e-34d2-410f-ac19-69a015e407b9 [ 641.878747] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 641.878747] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Starting heal instance info cache {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10257}} [ 641.878747] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Rebuilding the list of instances to heal {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10261}} [ 641.951408] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-820c5c4e-9077-4768-bf65-db8f20a6d92d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.959261] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bf5bde9-5ef3-4608-8b8b-2f032129a6f4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.989542] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48d803d8-4121-4b09-bc59-fb2934674cea {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.996841] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e132bde-594f-4ba5-9dcb-8fa8b8e0e6e1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.010892] env[62965]: DEBUG nova.compute.provider_tree [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 642.188489] env[62965]: DEBUG nova.network.neutron [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.253599] env[62965]: DEBUG oslo_concurrency.lockutils [None req-67dc2ac2-dfdf-410f-b90c-d2faf31877c7 tempest-ServerAddressesNegativeTestJSON-1869540618 tempest-ServerAddressesNegativeTestJSON-1869540618-project-member] Lock "53519d3e-34d2-410f-ac19-69a015e407b9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 81.828s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.382309] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Skipping network cache update for instance because it is Building. {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10270}} [ 642.382474] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Skipping network cache update for instance because it is Building. {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10270}} [ 642.382613] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Skipping network cache update for instance because it is Building. {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10270}} [ 642.382848] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Didn't find any instances for network info cache update. {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10343}} [ 642.382930] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 642.383089] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 642.383236] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 642.383380] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 642.383519] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 642.383658] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 642.383779] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62965) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10876}} [ 642.383920] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 642.516021] env[62965]: DEBUG nova.scheduler.client.report [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 642.692842] env[62965]: INFO nova.compute.manager [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] [instance: f594c50e-25b3-4e85-840c-7dfe0994a2c2] Took 1.05 seconds to deallocate network for instance. [ 642.756751] env[62965]: DEBUG nova.compute.manager [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 642.888136] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.018367] env[62965]: DEBUG oslo_concurrency.lockutils [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.499s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 643.018904] env[62965]: DEBUG nova.compute.manager [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 643.022050] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.005s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 643.288021] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.525474] env[62965]: DEBUG nova.compute.utils [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 643.525474] env[62965]: DEBUG nova.compute.manager [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 643.525474] env[62965]: DEBUG nova.network.neutron [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 643.628614] env[62965]: DEBUG nova.policy [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e5ccf1db0e914bd3afb47ceb37449f5f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c26a5af6390440cf8d6eb40ba101a9ba', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 643.745730] env[62965]: INFO nova.scheduler.client.report [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] Deleted allocations for instance f594c50e-25b3-4e85-840c-7dfe0994a2c2 [ 644.036278] env[62965]: DEBUG nova.compute.manager [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 644.044502] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78005499-452f-4af9-b5c7-5157b8f078c1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.054080] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b8b9114-b7d6-4a57-8021-6fd549b42e6d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.085051] env[62965]: DEBUG nova.network.neutron [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Successfully created port: d0654b21-e661-454c-8b61-f9b30be9a633 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 644.088404] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f0f8f72-3ad4-4d78-9fc2-7e91107761d7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.098098] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-714fc9a9-3409-4b4c-b82e-cc80f4c1ed8e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.113286] env[62965]: DEBUG nova.compute.provider_tree [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 644.260628] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d5757fe9-acec-4262-a05f-432d0c882529 tempest-ServersTestJSON-1688430543 tempest-ServersTestJSON-1688430543-project-member] Lock "f594c50e-25b3-4e85-840c-7dfe0994a2c2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 81.632s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.620008] env[62965]: DEBUG nova.scheduler.client.report [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 644.763107] env[62965]: DEBUG nova.compute.manager [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 645.057623] env[62965]: DEBUG nova.compute.manager [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 645.090091] env[62965]: DEBUG nova.virt.hardware [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 645.090347] env[62965]: DEBUG nova.virt.hardware [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 645.090512] env[62965]: DEBUG nova.virt.hardware [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 645.090728] env[62965]: DEBUG nova.virt.hardware [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 645.090920] env[62965]: DEBUG nova.virt.hardware [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 645.091491] env[62965]: DEBUG nova.virt.hardware [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 645.091857] env[62965]: DEBUG nova.virt.hardware [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 645.092521] env[62965]: DEBUG nova.virt.hardware [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 645.092909] env[62965]: DEBUG nova.virt.hardware [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 645.093192] env[62965]: DEBUG nova.virt.hardware [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 645.093471] env[62965]: DEBUG nova.virt.hardware [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 645.095848] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6944b9a-043a-414d-9288-077415b61efc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.107034] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-603c9e3d-928d-4008-b7f0-0bc76213563d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.127289] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.105s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.128387] env[62965]: ERROR nova.compute.manager [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b6c21631-dcbf-4d90-ad9b-fa7ddd03e087, please check neutron logs for more information. [ 645.128387] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Traceback (most recent call last): [ 645.128387] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 645.128387] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] self.driver.spawn(context, instance, image_meta, [ 645.128387] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 645.128387] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 645.128387] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 645.128387] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] vm_ref = self.build_virtual_machine(instance, [ 645.128387] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 645.128387] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] vif_infos = vmwarevif.get_vif_info(self._session, [ 645.128387] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 645.128716] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] for vif in network_info: [ 645.128716] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 645.128716] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] return self._sync_wrapper(fn, *args, **kwargs) [ 645.128716] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 645.128716] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] self.wait() [ 645.128716] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 645.128716] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] self[:] = self._gt.wait() [ 645.128716] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 645.128716] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] return self._exit_event.wait() [ 645.128716] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 645.128716] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] result = hub.switch() [ 645.128716] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 645.128716] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] return self.greenlet.switch() [ 645.129042] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 645.129042] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] result = function(*args, **kwargs) [ 645.129042] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 645.129042] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] return func(*args, **kwargs) [ 645.129042] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 645.129042] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] raise e [ 645.129042] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 645.129042] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] nwinfo = self.network_api.allocate_for_instance( [ 645.129042] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 645.129042] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] created_port_ids = self._update_ports_for_instance( [ 645.129042] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 645.129042] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] with excutils.save_and_reraise_exception(): [ 645.129042] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 645.129367] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] self.force_reraise() [ 645.129367] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 645.129367] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] raise self.value [ 645.129367] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 645.129367] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] updated_port = self._update_port( [ 645.129367] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 645.129367] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] _ensure_no_port_binding_failure(port) [ 645.129367] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 645.129367] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] raise exception.PortBindingFailed(port_id=port['id']) [ 645.129367] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] nova.exception.PortBindingFailed: Binding failed for port b6c21631-dcbf-4d90-ad9b-fa7ddd03e087, please check neutron logs for more information. [ 645.129367] env[62965]: ERROR nova.compute.manager [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] [ 645.129686] env[62965]: DEBUG nova.compute.utils [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Binding failed for port b6c21631-dcbf-4d90-ad9b-fa7ddd03e087, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 645.130645] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.779s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 645.132090] env[62965]: INFO nova.compute.claims [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 645.139230] env[62965]: DEBUG nova.compute.manager [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Build of instance 09b62788-3082-4c99-9dfc-0843c4988feb was re-scheduled: Binding failed for port b6c21631-dcbf-4d90-ad9b-fa7ddd03e087, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 645.139230] env[62965]: DEBUG nova.compute.manager [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 645.139230] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Acquiring lock "refresh_cache-09b62788-3082-4c99-9dfc-0843c4988feb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 645.139230] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Acquired lock "refresh_cache-09b62788-3082-4c99-9dfc-0843c4988feb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.139440] env[62965]: DEBUG nova.network.neutron [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 645.292386] env[62965]: DEBUG oslo_concurrency.lockutils [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.668862] env[62965]: DEBUG nova.network.neutron [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 645.749550] env[62965]: DEBUG nova.compute.manager [req-ff8bc68e-2718-4513-81d9-680d6c5ba4f6 req-d238f948-1091-410c-b365-85763786a2eb service nova] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Received event network-changed-d0654b21-e661-454c-8b61-f9b30be9a633 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 645.749738] env[62965]: DEBUG nova.compute.manager [req-ff8bc68e-2718-4513-81d9-680d6c5ba4f6 req-d238f948-1091-410c-b365-85763786a2eb service nova] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Refreshing instance network info cache due to event network-changed-d0654b21-e661-454c-8b61-f9b30be9a633. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 645.749943] env[62965]: DEBUG oslo_concurrency.lockutils [req-ff8bc68e-2718-4513-81d9-680d6c5ba4f6 req-d238f948-1091-410c-b365-85763786a2eb service nova] Acquiring lock "refresh_cache-222f547e-8fce-4576-bafc-809ea2ee8631" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 645.750096] env[62965]: DEBUG oslo_concurrency.lockutils [req-ff8bc68e-2718-4513-81d9-680d6c5ba4f6 req-d238f948-1091-410c-b365-85763786a2eb service nova] Acquired lock "refresh_cache-222f547e-8fce-4576-bafc-809ea2ee8631" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.750280] env[62965]: DEBUG nova.network.neutron [req-ff8bc68e-2718-4513-81d9-680d6c5ba4f6 req-d238f948-1091-410c-b365-85763786a2eb service nova] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Refreshing network info cache for port d0654b21-e661-454c-8b61-f9b30be9a633 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 646.082979] env[62965]: ERROR nova.compute.manager [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d0654b21-e661-454c-8b61-f9b30be9a633, please check neutron logs for more information. [ 646.082979] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 646.082979] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 646.082979] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 646.082979] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 646.082979] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 646.082979] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 646.082979] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 646.082979] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 646.082979] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 646.082979] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 646.082979] env[62965]: ERROR nova.compute.manager raise self.value [ 646.082979] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 646.082979] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 646.082979] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 646.082979] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 646.083579] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 646.083579] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 646.083579] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d0654b21-e661-454c-8b61-f9b30be9a633, please check neutron logs for more information. [ 646.083579] env[62965]: ERROR nova.compute.manager [ 646.083579] env[62965]: Traceback (most recent call last): [ 646.083579] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 646.083579] env[62965]: listener.cb(fileno) [ 646.083579] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 646.083579] env[62965]: result = function(*args, **kwargs) [ 646.083579] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 646.083579] env[62965]: return func(*args, **kwargs) [ 646.083579] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 646.083579] env[62965]: raise e [ 646.083579] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 646.083579] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 646.083579] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 646.083579] env[62965]: created_port_ids = self._update_ports_for_instance( [ 646.083579] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 646.083579] env[62965]: with excutils.save_and_reraise_exception(): [ 646.083579] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 646.083579] env[62965]: self.force_reraise() [ 646.083579] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 646.083579] env[62965]: raise self.value [ 646.083579] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 646.083579] env[62965]: updated_port = self._update_port( [ 646.083579] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 646.083579] env[62965]: _ensure_no_port_binding_failure(port) [ 646.083579] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 646.083579] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 646.084256] env[62965]: nova.exception.PortBindingFailed: Binding failed for port d0654b21-e661-454c-8b61-f9b30be9a633, please check neutron logs for more information. [ 646.084256] env[62965]: Removing descriptor: 19 [ 646.084256] env[62965]: ERROR nova.compute.manager [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d0654b21-e661-454c-8b61-f9b30be9a633, please check neutron logs for more information. [ 646.084256] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Traceback (most recent call last): [ 646.084256] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 646.084256] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] yield resources [ 646.084256] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 646.084256] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] self.driver.spawn(context, instance, image_meta, [ 646.084256] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 646.084256] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] self._vmops.spawn(context, instance, image_meta, injected_files, [ 646.084256] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 646.084256] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] vm_ref = self.build_virtual_machine(instance, [ 646.084551] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 646.084551] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] vif_infos = vmwarevif.get_vif_info(self._session, [ 646.084551] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 646.084551] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] for vif in network_info: [ 646.084551] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 646.084551] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] return self._sync_wrapper(fn, *args, **kwargs) [ 646.084551] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 646.084551] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] self.wait() [ 646.084551] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 646.084551] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] self[:] = self._gt.wait() [ 646.084551] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 646.084551] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] return self._exit_event.wait() [ 646.084551] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 646.084911] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] result = hub.switch() [ 646.084911] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 646.084911] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] return self.greenlet.switch() [ 646.084911] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 646.084911] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] result = function(*args, **kwargs) [ 646.084911] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 646.084911] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] return func(*args, **kwargs) [ 646.084911] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 646.084911] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] raise e [ 646.084911] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 646.084911] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] nwinfo = self.network_api.allocate_for_instance( [ 646.084911] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 646.084911] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] created_port_ids = self._update_ports_for_instance( [ 646.085211] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 646.085211] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] with excutils.save_and_reraise_exception(): [ 646.085211] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 646.085211] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] self.force_reraise() [ 646.085211] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 646.085211] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] raise self.value [ 646.085211] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 646.085211] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] updated_port = self._update_port( [ 646.085211] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 646.085211] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] _ensure_no_port_binding_failure(port) [ 646.085211] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 646.085211] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] raise exception.PortBindingFailed(port_id=port['id']) [ 646.085487] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] nova.exception.PortBindingFailed: Binding failed for port d0654b21-e661-454c-8b61-f9b30be9a633, please check neutron logs for more information. [ 646.085487] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] [ 646.085487] env[62965]: INFO nova.compute.manager [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Terminating instance [ 646.096267] env[62965]: DEBUG nova.network.neutron [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.288826] env[62965]: DEBUG nova.network.neutron [req-ff8bc68e-2718-4513-81d9-680d6c5ba4f6 req-d238f948-1091-410c-b365-85763786a2eb service nova] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 646.541337] env[62965]: DEBUG nova.network.neutron [req-ff8bc68e-2718-4513-81d9-680d6c5ba4f6 req-d238f948-1091-410c-b365-85763786a2eb service nova] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.591198] env[62965]: DEBUG oslo_concurrency.lockutils [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] Acquiring lock "refresh_cache-222f547e-8fce-4576-bafc-809ea2ee8631" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 646.599552] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Releasing lock "refresh_cache-09b62788-3082-4c99-9dfc-0843c4988feb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 646.599717] env[62965]: DEBUG nova.compute.manager [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 646.599903] env[62965]: DEBUG nova.compute.manager [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 646.600108] env[62965]: DEBUG nova.network.neutron [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 646.621659] env[62965]: DEBUG nova.network.neutron [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 646.638773] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7a72973-d21a-477c-9d2a-beba2a263a71 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.647140] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8771e75d-75bf-4504-9853-bb614f70de70 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.681996] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-125fafc5-ef88-480b-8cf8-17f9af3b6604 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.691351] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b31991cd-b3f4-49a8-b942-3ec4f1e2a3ad {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.706820] env[62965]: DEBUG nova.compute.provider_tree [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 647.045285] env[62965]: DEBUG oslo_concurrency.lockutils [req-ff8bc68e-2718-4513-81d9-680d6c5ba4f6 req-d238f948-1091-410c-b365-85763786a2eb service nova] Releasing lock "refresh_cache-222f547e-8fce-4576-bafc-809ea2ee8631" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.045285] env[62965]: DEBUG oslo_concurrency.lockutils [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] Acquired lock "refresh_cache-222f547e-8fce-4576-bafc-809ea2ee8631" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.046236] env[62965]: DEBUG nova.network.neutron [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 647.131724] env[62965]: DEBUG nova.network.neutron [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.211103] env[62965]: DEBUG nova.scheduler.client.report [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 647.572972] env[62965]: DEBUG nova.network.neutron [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 647.638302] env[62965]: INFO nova.compute.manager [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 09b62788-3082-4c99-9dfc-0843c4988feb] Took 1.04 seconds to deallocate network for instance. [ 647.720090] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.589s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.720609] env[62965]: DEBUG nova.compute.manager [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 647.723131] env[62965]: DEBUG oslo_concurrency.lockutils [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.378s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.724777] env[62965]: INFO nova.compute.claims [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 647.735156] env[62965]: DEBUG nova.network.neutron [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.827560] env[62965]: DEBUG nova.compute.manager [req-323ad58f-913d-469b-af49-ceaf2bb88bcb req-4382eb4e-5ef7-416d-aa91-630c6859c484 service nova] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Received event network-vif-deleted-d0654b21-e661-454c-8b61-f9b30be9a633 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 648.229574] env[62965]: DEBUG nova.compute.utils [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 648.238927] env[62965]: DEBUG nova.compute.manager [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 648.238927] env[62965]: DEBUG nova.network.neutron [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 648.243160] env[62965]: DEBUG oslo_concurrency.lockutils [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] Releasing lock "refresh_cache-222f547e-8fce-4576-bafc-809ea2ee8631" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 648.243160] env[62965]: DEBUG nova.compute.manager [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 648.243160] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 648.243546] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-763eefdd-1d49-4ba4-9de1-176ec3b10f2d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.258933] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1055d038-fe81-4cf0-b15f-6286dacc7940 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.284443] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 222f547e-8fce-4576-bafc-809ea2ee8631 could not be found. [ 648.287257] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 648.287257] env[62965]: INFO nova.compute.manager [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Took 0.04 seconds to destroy the instance on the hypervisor. [ 648.287257] env[62965]: DEBUG oslo.service.loopingcall [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 648.287257] env[62965]: DEBUG nova.compute.manager [-] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 648.287257] env[62965]: DEBUG nova.network.neutron [-] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 648.308561] env[62965]: DEBUG nova.network.neutron [-] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 648.361122] env[62965]: DEBUG nova.policy [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7dca5bc95b22405b8f04af2808277009', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '346df53fb497435b9e0c1157c502e8e9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 648.671912] env[62965]: INFO nova.scheduler.client.report [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Deleted allocations for instance 09b62788-3082-4c99-9dfc-0843c4988feb [ 648.737436] env[62965]: DEBUG nova.compute.manager [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 648.811628] env[62965]: DEBUG nova.network.neutron [-] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.165615] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94d07001-2953-4414-968f-1824a0381036 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.173609] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86c20632-f92a-419f-9e3f-35047ba23c77 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.210821] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e8be3ae7-c87d-4615-942b-8b4e387441ce tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Lock "09b62788-3082-4c99-9dfc-0843c4988feb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 85.833s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 649.212593] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b188f4e-0ade-4106-ab08-9e198475560d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.221365] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f2f6c8b-8579-4b03-868d-1c6f33de9c03 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.239425] env[62965]: DEBUG nova.compute.provider_tree [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 649.275132] env[62965]: DEBUG nova.network.neutron [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Successfully created port: f0cea45c-6193-4d2f-8a27-0fee3faf6256 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 649.314147] env[62965]: INFO nova.compute.manager [-] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Took 1.03 seconds to deallocate network for instance. [ 649.317617] env[62965]: DEBUG nova.compute.claims [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 649.318390] env[62965]: DEBUG oslo_concurrency.lockutils [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.719506] env[62965]: DEBUG nova.compute.manager [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 649.742507] env[62965]: DEBUG nova.scheduler.client.report [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 649.756511] env[62965]: DEBUG nova.compute.manager [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 649.791724] env[62965]: DEBUG nova.virt.hardware [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 649.791958] env[62965]: DEBUG nova.virt.hardware [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 649.792132] env[62965]: DEBUG nova.virt.hardware [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 649.792316] env[62965]: DEBUG nova.virt.hardware [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 649.792559] env[62965]: DEBUG nova.virt.hardware [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 649.792616] env[62965]: DEBUG nova.virt.hardware [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 649.792785] env[62965]: DEBUG nova.virt.hardware [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 649.792935] env[62965]: DEBUG nova.virt.hardware [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 649.798163] env[62965]: DEBUG nova.virt.hardware [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 649.798163] env[62965]: DEBUG nova.virt.hardware [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 649.798163] env[62965]: DEBUG nova.virt.hardware [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 649.800353] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6946728d-0c25-4f0e-bdc7-fd5c183dd43a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.809894] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b4498ac-2df3-4062-9db0-e60a6965be8a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.249126] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 650.252030] env[62965]: DEBUG oslo_concurrency.lockutils [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.529s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 650.252516] env[62965]: DEBUG nova.compute.manager [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 650.255827] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.694s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 650.446076] env[62965]: DEBUG oslo_concurrency.lockutils [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Acquiring lock "6e0a4e4b-21a2-4185-999e-69b4c74c37d7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 650.446076] env[62965]: DEBUG oslo_concurrency.lockutils [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Lock "6e0a4e4b-21a2-4185-999e-69b4c74c37d7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 650.762059] env[62965]: DEBUG nova.compute.utils [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 650.770722] env[62965]: DEBUG nova.compute.manager [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 650.771791] env[62965]: DEBUG nova.network.neutron [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 650.774596] env[62965]: DEBUG nova.compute.manager [req-d05b79dd-23f8-4420-a50b-63606e1c0366 req-1289308a-4794-4862-b6a3-7f42319ffc0f service nova] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Received event network-changed-f0cea45c-6193-4d2f-8a27-0fee3faf6256 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 650.774769] env[62965]: DEBUG nova.compute.manager [req-d05b79dd-23f8-4420-a50b-63606e1c0366 req-1289308a-4794-4862-b6a3-7f42319ffc0f service nova] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Refreshing instance network info cache due to event network-changed-f0cea45c-6193-4d2f-8a27-0fee3faf6256. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 650.776083] env[62965]: DEBUG oslo_concurrency.lockutils [req-d05b79dd-23f8-4420-a50b-63606e1c0366 req-1289308a-4794-4862-b6a3-7f42319ffc0f service nova] Acquiring lock "refresh_cache-af29af38-bd31-446a-a944-2bb43d9aa6c6" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 650.776083] env[62965]: DEBUG oslo_concurrency.lockutils [req-d05b79dd-23f8-4420-a50b-63606e1c0366 req-1289308a-4794-4862-b6a3-7f42319ffc0f service nova] Acquired lock "refresh_cache-af29af38-bd31-446a-a944-2bb43d9aa6c6" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.776083] env[62965]: DEBUG nova.network.neutron [req-d05b79dd-23f8-4420-a50b-63606e1c0366 req-1289308a-4794-4862-b6a3-7f42319ffc0f service nova] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Refreshing network info cache for port f0cea45c-6193-4d2f-8a27-0fee3faf6256 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 650.837295] env[62965]: DEBUG nova.policy [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b586afcd74b34c36b03dd6f35f2e962b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e1013f4d0e864d3f9630f65f2f9a643e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 650.865483] env[62965]: ERROR nova.compute.manager [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f0cea45c-6193-4d2f-8a27-0fee3faf6256, please check neutron logs for more information. [ 650.865483] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 650.865483] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 650.865483] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 650.865483] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 650.865483] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 650.865483] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 650.865483] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 650.865483] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 650.865483] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 650.865483] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 650.865483] env[62965]: ERROR nova.compute.manager raise self.value [ 650.865483] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 650.865483] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 650.865483] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 650.865483] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 650.866056] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 650.866056] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 650.866056] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f0cea45c-6193-4d2f-8a27-0fee3faf6256, please check neutron logs for more information. [ 650.866056] env[62965]: ERROR nova.compute.manager [ 650.866056] env[62965]: Traceback (most recent call last): [ 650.866056] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 650.866056] env[62965]: listener.cb(fileno) [ 650.866056] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 650.866056] env[62965]: result = function(*args, **kwargs) [ 650.866056] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 650.866056] env[62965]: return func(*args, **kwargs) [ 650.866056] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 650.866056] env[62965]: raise e [ 650.866056] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 650.866056] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 650.866056] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 650.866056] env[62965]: created_port_ids = self._update_ports_for_instance( [ 650.866056] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 650.866056] env[62965]: with excutils.save_and_reraise_exception(): [ 650.866056] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 650.866056] env[62965]: self.force_reraise() [ 650.866056] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 650.866056] env[62965]: raise self.value [ 650.866056] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 650.866056] env[62965]: updated_port = self._update_port( [ 650.866056] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 650.866056] env[62965]: _ensure_no_port_binding_failure(port) [ 650.866056] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 650.866056] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 650.866861] env[62965]: nova.exception.PortBindingFailed: Binding failed for port f0cea45c-6193-4d2f-8a27-0fee3faf6256, please check neutron logs for more information. [ 650.866861] env[62965]: Removing descriptor: 17 [ 650.866861] env[62965]: ERROR nova.compute.manager [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f0cea45c-6193-4d2f-8a27-0fee3faf6256, please check neutron logs for more information. [ 650.866861] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Traceback (most recent call last): [ 650.866861] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 650.866861] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] yield resources [ 650.866861] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 650.866861] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] self.driver.spawn(context, instance, image_meta, [ 650.866861] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 650.866861] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 650.866861] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 650.866861] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] vm_ref = self.build_virtual_machine(instance, [ 650.867166] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 650.867166] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] vif_infos = vmwarevif.get_vif_info(self._session, [ 650.867166] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 650.867166] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] for vif in network_info: [ 650.867166] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 650.867166] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] return self._sync_wrapper(fn, *args, **kwargs) [ 650.867166] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 650.867166] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] self.wait() [ 650.867166] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 650.867166] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] self[:] = self._gt.wait() [ 650.867166] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 650.867166] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] return self._exit_event.wait() [ 650.867166] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 650.870315] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] result = hub.switch() [ 650.870315] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 650.870315] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] return self.greenlet.switch() [ 650.870315] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 650.870315] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] result = function(*args, **kwargs) [ 650.870315] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 650.870315] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] return func(*args, **kwargs) [ 650.870315] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 650.870315] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] raise e [ 650.870315] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 650.870315] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] nwinfo = self.network_api.allocate_for_instance( [ 650.870315] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 650.870315] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] created_port_ids = self._update_ports_for_instance( [ 650.870882] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 650.870882] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] with excutils.save_and_reraise_exception(): [ 650.870882] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 650.870882] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] self.force_reraise() [ 650.870882] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 650.870882] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] raise self.value [ 650.870882] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 650.870882] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] updated_port = self._update_port( [ 650.870882] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 650.870882] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] _ensure_no_port_binding_failure(port) [ 650.870882] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 650.870882] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] raise exception.PortBindingFailed(port_id=port['id']) [ 650.872623] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] nova.exception.PortBindingFailed: Binding failed for port f0cea45c-6193-4d2f-8a27-0fee3faf6256, please check neutron logs for more information. [ 650.872623] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] [ 650.872623] env[62965]: INFO nova.compute.manager [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Terminating instance [ 650.988348] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Acquiring lock "e87adfae-3ea4-4cc4-9978-d8c811c80941" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 650.989758] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Lock "e87adfae-3ea4-4cc4-9978-d8c811c80941" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.272307] env[62965]: DEBUG nova.compute.manager [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 651.302516] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bb759a0-dae7-4f86-9ca8-23be0d8d300f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.316960] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b34cd6fd-ec4b-4dc3-917d-c1b41daf565e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.349666] env[62965]: DEBUG nova.network.neutron [req-d05b79dd-23f8-4420-a50b-63606e1c0366 req-1289308a-4794-4862-b6a3-7f42319ffc0f service nova] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 651.356638] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe67a74c-6640-4d72-9ff6-b0a3ff66befc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.367370] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7d1a568-dee7-4ce4-a1f7-477057480bff {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.383836] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Acquiring lock "refresh_cache-af29af38-bd31-446a-a944-2bb43d9aa6c6" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.384752] env[62965]: DEBUG nova.compute.provider_tree [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 651.529191] env[62965]: DEBUG nova.network.neutron [req-d05b79dd-23f8-4420-a50b-63606e1c0366 req-1289308a-4794-4862-b6a3-7f42319ffc0f service nova] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.555534] env[62965]: DEBUG nova.network.neutron [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Successfully created port: 6764b048-89fb-436a-a183-ce56c1079e65 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 651.891622] env[62965]: DEBUG nova.scheduler.client.report [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 651.912844] env[62965]: DEBUG oslo_concurrency.lockutils [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Acquiring lock "9a2a324c-4eb1-45b6-8414-42671b412429" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.912844] env[62965]: DEBUG oslo_concurrency.lockutils [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Lock "9a2a324c-4eb1-45b6-8414-42671b412429" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.033495] env[62965]: DEBUG oslo_concurrency.lockutils [req-d05b79dd-23f8-4420-a50b-63606e1c0366 req-1289308a-4794-4862-b6a3-7f42319ffc0f service nova] Releasing lock "refresh_cache-af29af38-bd31-446a-a944-2bb43d9aa6c6" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 652.033495] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Acquired lock "refresh_cache-af29af38-bd31-446a-a944-2bb43d9aa6c6" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.033495] env[62965]: DEBUG nova.network.neutron [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 652.287876] env[62965]: DEBUG nova.compute.manager [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 652.329291] env[62965]: DEBUG nova.virt.hardware [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 652.329561] env[62965]: DEBUG nova.virt.hardware [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 652.329819] env[62965]: DEBUG nova.virt.hardware [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 652.329888] env[62965]: DEBUG nova.virt.hardware [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 652.330074] env[62965]: DEBUG nova.virt.hardware [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 652.330250] env[62965]: DEBUG nova.virt.hardware [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 652.330438] env[62965]: DEBUG nova.virt.hardware [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 652.330601] env[62965]: DEBUG nova.virt.hardware [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 652.330769] env[62965]: DEBUG nova.virt.hardware [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 652.330986] env[62965]: DEBUG nova.virt.hardware [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 652.331144] env[62965]: DEBUG nova.virt.hardware [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 652.332323] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5720096-a816-4d2f-8933-8c232c44542f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.343601] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-703ca0d6-0378-4306-a8db-bd5f75e58cf0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.400900] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.145s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 652.401300] env[62965]: ERROR nova.compute.manager [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c768e9b3-53f1-42b5-b16e-edd046109356, please check neutron logs for more information. [ 652.401300] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Traceback (most recent call last): [ 652.401300] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 652.401300] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] self.driver.spawn(context, instance, image_meta, [ 652.401300] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 652.401300] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 652.401300] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 652.401300] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] vm_ref = self.build_virtual_machine(instance, [ 652.401300] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 652.401300] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] vif_infos = vmwarevif.get_vif_info(self._session, [ 652.401300] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 652.401611] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] for vif in network_info: [ 652.401611] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 652.401611] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] return self._sync_wrapper(fn, *args, **kwargs) [ 652.401611] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 652.401611] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] self.wait() [ 652.401611] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 652.401611] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] self[:] = self._gt.wait() [ 652.401611] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 652.401611] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] return self._exit_event.wait() [ 652.401611] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 652.401611] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] result = hub.switch() [ 652.401611] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 652.401611] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] return self.greenlet.switch() [ 652.402100] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 652.402100] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] result = function(*args, **kwargs) [ 652.402100] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 652.402100] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] return func(*args, **kwargs) [ 652.402100] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 652.402100] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] raise e [ 652.402100] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 652.402100] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] nwinfo = self.network_api.allocate_for_instance( [ 652.402100] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 652.402100] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] created_port_ids = self._update_ports_for_instance( [ 652.402100] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 652.402100] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] with excutils.save_and_reraise_exception(): [ 652.402100] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 652.402488] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] self.force_reraise() [ 652.402488] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 652.402488] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] raise self.value [ 652.402488] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 652.402488] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] updated_port = self._update_port( [ 652.402488] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 652.402488] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] _ensure_no_port_binding_failure(port) [ 652.402488] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 652.402488] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] raise exception.PortBindingFailed(port_id=port['id']) [ 652.402488] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] nova.exception.PortBindingFailed: Binding failed for port c768e9b3-53f1-42b5-b16e-edd046109356, please check neutron logs for more information. [ 652.402488] env[62965]: ERROR nova.compute.manager [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] [ 652.402783] env[62965]: DEBUG nova.compute.utils [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Binding failed for port c768e9b3-53f1-42b5-b16e-edd046109356, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 652.405971] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.245s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.406848] env[62965]: INFO nova.compute.claims [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 652.410666] env[62965]: DEBUG nova.compute.manager [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Build of instance 63c54ed4-4525-42f6-a99f-033f7b9838a7 was re-scheduled: Binding failed for port c768e9b3-53f1-42b5-b16e-edd046109356, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 652.412190] env[62965]: DEBUG nova.compute.manager [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 652.412190] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] Acquiring lock "refresh_cache-63c54ed4-4525-42f6-a99f-033f7b9838a7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 652.412190] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] Acquired lock "refresh_cache-63c54ed4-4525-42f6-a99f-033f7b9838a7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.412190] env[62965]: DEBUG nova.network.neutron [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 652.565581] env[62965]: DEBUG nova.network.neutron [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 652.698445] env[62965]: DEBUG nova.network.neutron [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.948435] env[62965]: DEBUG nova.network.neutron [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 652.989833] env[62965]: DEBUG nova.compute.manager [req-09b0cf62-acb9-493c-995d-1fa1b870c910 req-2f3bc493-0d00-4e96-aab5-4663f69aa119 service nova] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Received event network-vif-deleted-f0cea45c-6193-4d2f-8a27-0fee3faf6256 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 653.057074] env[62965]: DEBUG nova.network.neutron [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.153074] env[62965]: DEBUG nova.compute.manager [req-ae8b6338-7cc9-470f-8c3f-f768bf7fc10e req-5d30e8bb-4474-4b1a-b6dd-8b2eca959717 service nova] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Received event network-changed-6764b048-89fb-436a-a183-ce56c1079e65 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 653.153074] env[62965]: DEBUG nova.compute.manager [req-ae8b6338-7cc9-470f-8c3f-f768bf7fc10e req-5d30e8bb-4474-4b1a-b6dd-8b2eca959717 service nova] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Refreshing instance network info cache due to event network-changed-6764b048-89fb-436a-a183-ce56c1079e65. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 653.153074] env[62965]: DEBUG oslo_concurrency.lockutils [req-ae8b6338-7cc9-470f-8c3f-f768bf7fc10e req-5d30e8bb-4474-4b1a-b6dd-8b2eca959717 service nova] Acquiring lock "refresh_cache-8a893e6e-37d4-4fd4-b0db-d6633117bdce" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 653.153074] env[62965]: DEBUG oslo_concurrency.lockutils [req-ae8b6338-7cc9-470f-8c3f-f768bf7fc10e req-5d30e8bb-4474-4b1a-b6dd-8b2eca959717 service nova] Acquired lock "refresh_cache-8a893e6e-37d4-4fd4-b0db-d6633117bdce" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.153074] env[62965]: DEBUG nova.network.neutron [req-ae8b6338-7cc9-470f-8c3f-f768bf7fc10e req-5d30e8bb-4474-4b1a-b6dd-8b2eca959717 service nova] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Refreshing network info cache for port 6764b048-89fb-436a-a183-ce56c1079e65 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 653.203948] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Releasing lock "refresh_cache-af29af38-bd31-446a-a944-2bb43d9aa6c6" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 653.204432] env[62965]: DEBUG nova.compute.manager [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 653.204627] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 653.205025] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7f3e51ad-a2b0-41b6-a798-13715c61f5ce {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.214294] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a26452d-7cef-41ef-8695-9ea831a4455a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.242542] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance af29af38-bd31-446a-a944-2bb43d9aa6c6 could not be found. [ 653.242542] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 653.246105] env[62965]: INFO nova.compute.manager [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Took 0.04 seconds to destroy the instance on the hypervisor. [ 653.246105] env[62965]: DEBUG oslo.service.loopingcall [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 653.246105] env[62965]: DEBUG nova.compute.manager [-] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 653.246105] env[62965]: DEBUG nova.network.neutron [-] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 653.268451] env[62965]: DEBUG nova.network.neutron [-] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 653.341637] env[62965]: ERROR nova.compute.manager [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6764b048-89fb-436a-a183-ce56c1079e65, please check neutron logs for more information. [ 653.341637] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 653.341637] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 653.341637] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 653.341637] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 653.341637] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 653.341637] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 653.341637] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 653.341637] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 653.341637] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 653.341637] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 653.341637] env[62965]: ERROR nova.compute.manager raise self.value [ 653.341637] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 653.341637] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 653.341637] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 653.341637] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 653.342242] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 653.342242] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 653.342242] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6764b048-89fb-436a-a183-ce56c1079e65, please check neutron logs for more information. [ 653.342242] env[62965]: ERROR nova.compute.manager [ 653.342242] env[62965]: Traceback (most recent call last): [ 653.342242] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 653.342242] env[62965]: listener.cb(fileno) [ 653.342242] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 653.342242] env[62965]: result = function(*args, **kwargs) [ 653.342242] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 653.342242] env[62965]: return func(*args, **kwargs) [ 653.342242] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 653.342242] env[62965]: raise e [ 653.342242] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 653.342242] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 653.342242] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 653.342242] env[62965]: created_port_ids = self._update_ports_for_instance( [ 653.342242] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 653.342242] env[62965]: with excutils.save_and_reraise_exception(): [ 653.342242] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 653.342242] env[62965]: self.force_reraise() [ 653.342242] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 653.342242] env[62965]: raise self.value [ 653.342242] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 653.342242] env[62965]: updated_port = self._update_port( [ 653.342242] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 653.342242] env[62965]: _ensure_no_port_binding_failure(port) [ 653.342242] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 653.342242] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 653.342931] env[62965]: nova.exception.PortBindingFailed: Binding failed for port 6764b048-89fb-436a-a183-ce56c1079e65, please check neutron logs for more information. [ 653.342931] env[62965]: Removing descriptor: 19 [ 653.342931] env[62965]: ERROR nova.compute.manager [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6764b048-89fb-436a-a183-ce56c1079e65, please check neutron logs for more information. [ 653.342931] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Traceback (most recent call last): [ 653.342931] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 653.342931] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] yield resources [ 653.342931] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 653.342931] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] self.driver.spawn(context, instance, image_meta, [ 653.342931] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 653.342931] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] self._vmops.spawn(context, instance, image_meta, injected_files, [ 653.342931] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 653.342931] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] vm_ref = self.build_virtual_machine(instance, [ 653.343239] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 653.343239] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] vif_infos = vmwarevif.get_vif_info(self._session, [ 653.343239] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 653.343239] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] for vif in network_info: [ 653.343239] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 653.343239] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] return self._sync_wrapper(fn, *args, **kwargs) [ 653.343239] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 653.343239] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] self.wait() [ 653.343239] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 653.343239] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] self[:] = self._gt.wait() [ 653.343239] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 653.343239] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] return self._exit_event.wait() [ 653.343239] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 653.343550] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] result = hub.switch() [ 653.343550] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 653.343550] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] return self.greenlet.switch() [ 653.343550] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 653.343550] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] result = function(*args, **kwargs) [ 653.343550] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 653.343550] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] return func(*args, **kwargs) [ 653.343550] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 653.343550] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] raise e [ 653.343550] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 653.343550] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] nwinfo = self.network_api.allocate_for_instance( [ 653.343550] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 653.343550] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] created_port_ids = self._update_ports_for_instance( [ 653.345323] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 653.345323] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] with excutils.save_and_reraise_exception(): [ 653.345323] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 653.345323] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] self.force_reraise() [ 653.345323] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 653.345323] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] raise self.value [ 653.345323] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 653.345323] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] updated_port = self._update_port( [ 653.345323] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 653.345323] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] _ensure_no_port_binding_failure(port) [ 653.345323] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 653.345323] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] raise exception.PortBindingFailed(port_id=port['id']) [ 653.345913] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] nova.exception.PortBindingFailed: Binding failed for port 6764b048-89fb-436a-a183-ce56c1079e65, please check neutron logs for more information. [ 653.345913] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] [ 653.345913] env[62965]: INFO nova.compute.manager [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Terminating instance [ 653.562628] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] Releasing lock "refresh_cache-63c54ed4-4525-42f6-a99f-033f7b9838a7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 653.562947] env[62965]: DEBUG nova.compute.manager [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 653.563193] env[62965]: DEBUG nova.compute.manager [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 653.563481] env[62965]: DEBUG nova.network.neutron [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 653.591236] env[62965]: DEBUG nova.network.neutron [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 653.674644] env[62965]: DEBUG nova.network.neutron [req-ae8b6338-7cc9-470f-8c3f-f768bf7fc10e req-5d30e8bb-4474-4b1a-b6dd-8b2eca959717 service nova] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 653.764878] env[62965]: DEBUG nova.network.neutron [req-ae8b6338-7cc9-470f-8c3f-f768bf7fc10e req-5d30e8bb-4474-4b1a-b6dd-8b2eca959717 service nova] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.772020] env[62965]: DEBUG nova.network.neutron [-] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.846993] env[62965]: DEBUG oslo_concurrency.lockutils [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Acquiring lock "refresh_cache-8a893e6e-37d4-4fd4-b0db-d6633117bdce" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 653.873278] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-281fe1df-4399-41aa-a968-e29b02233ab3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.880957] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0f5e508-2e5d-4ac8-900b-bc8dbb0504a2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.912591] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cb6a9ca-bf04-436f-a467-4992861e0e09 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.922009] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7636dea8-012c-4a9c-b7f7-af34ddcc36a9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.938102] env[62965]: DEBUG nova.compute.provider_tree [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 654.095653] env[62965]: DEBUG nova.network.neutron [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.267688] env[62965]: DEBUG oslo_concurrency.lockutils [req-ae8b6338-7cc9-470f-8c3f-f768bf7fc10e req-5d30e8bb-4474-4b1a-b6dd-8b2eca959717 service nova] Releasing lock "refresh_cache-8a893e6e-37d4-4fd4-b0db-d6633117bdce" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 654.268119] env[62965]: DEBUG oslo_concurrency.lockutils [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Acquired lock "refresh_cache-8a893e6e-37d4-4fd4-b0db-d6633117bdce" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.268358] env[62965]: DEBUG nova.network.neutron [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 654.274132] env[62965]: INFO nova.compute.manager [-] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Took 1.03 seconds to deallocate network for instance. [ 654.276260] env[62965]: DEBUG nova.compute.claims [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 654.276430] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 654.441622] env[62965]: DEBUG nova.scheduler.client.report [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 654.597281] env[62965]: INFO nova.compute.manager [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] [instance: 63c54ed4-4525-42f6-a99f-033f7b9838a7] Took 1.03 seconds to deallocate network for instance. [ 654.790663] env[62965]: DEBUG nova.network.neutron [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 654.865423] env[62965]: DEBUG nova.network.neutron [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.947932] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.543s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 654.948546] env[62965]: DEBUG nova.compute.manager [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 654.951551] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.771s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 654.952936] env[62965]: INFO nova.compute.claims [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 655.354951] env[62965]: DEBUG nova.compute.manager [req-6610177d-522b-4457-a447-19ce252e36ee req-6c1cc5a4-d402-46bc-86da-dced08ba2a5a service nova] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Received event network-vif-deleted-6764b048-89fb-436a-a183-ce56c1079e65 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 655.368592] env[62965]: DEBUG oslo_concurrency.lockutils [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Releasing lock "refresh_cache-8a893e6e-37d4-4fd4-b0db-d6633117bdce" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 655.369141] env[62965]: DEBUG nova.compute.manager [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 655.369420] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 655.369724] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3a906db8-3ac7-459a-a78b-5228a159a793 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.379961] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ada4e896-8f20-4f09-97fb-1e81ca18cd90 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.407532] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8a893e6e-37d4-4fd4-b0db-d6633117bdce could not be found. [ 655.407806] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 655.408020] env[62965]: INFO nova.compute.manager [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Took 0.04 seconds to destroy the instance on the hypervisor. [ 655.408292] env[62965]: DEBUG oslo.service.loopingcall [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 655.408575] env[62965]: DEBUG nova.compute.manager [-] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 655.408682] env[62965]: DEBUG nova.network.neutron [-] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 655.423627] env[62965]: DEBUG nova.network.neutron [-] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.457154] env[62965]: DEBUG nova.compute.utils [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 655.460553] env[62965]: DEBUG nova.compute.manager [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 655.460908] env[62965]: DEBUG nova.network.neutron [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 655.501197] env[62965]: DEBUG nova.policy [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '06ce81ccb906488398ac5e038ff786f5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '62caf44d43074f99aae93c87a6fead1c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 655.633721] env[62965]: INFO nova.scheduler.client.report [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] Deleted allocations for instance 63c54ed4-4525-42f6-a99f-033f7b9838a7 [ 655.772303] env[62965]: DEBUG nova.network.neutron [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Successfully created port: 763d2a98-a632-43dd-8f7e-f3c96720096e {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 655.928324] env[62965]: DEBUG nova.network.neutron [-] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.964492] env[62965]: DEBUG nova.compute.manager [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 656.146157] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1ba0d252-532c-443d-a2e9-75c7d66b358f tempest-AttachInterfacesUnderV243Test-906941893 tempest-AttachInterfacesUnderV243Test-906941893-project-member] Lock "63c54ed4-4525-42f6-a99f-033f7b9838a7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 92.227s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.418847] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f3f96df-3d6a-4e35-82c2-5d891d1837d0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.430967] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e028d1d-1920-4b96-b235-7f2b8af4a371 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.436145] env[62965]: INFO nova.compute.manager [-] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Took 1.03 seconds to deallocate network for instance. [ 656.438538] env[62965]: DEBUG nova.compute.claims [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 656.438804] env[62965]: DEBUG oslo_concurrency.lockutils [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.471557] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0d76593-479f-4b8c-8747-6e9633999275 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.487274] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28b914f5-313d-4452-aaee-3fd0a867e938 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.497921] env[62965]: DEBUG nova.compute.provider_tree [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 656.651104] env[62965]: DEBUG nova.compute.manager [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 656.981099] env[62965]: DEBUG nova.compute.manager [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 657.002586] env[62965]: DEBUG nova.scheduler.client.report [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 657.023488] env[62965]: DEBUG nova.virt.hardware [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 657.023772] env[62965]: DEBUG nova.virt.hardware [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 657.023929] env[62965]: DEBUG nova.virt.hardware [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 657.024131] env[62965]: DEBUG nova.virt.hardware [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 657.024280] env[62965]: DEBUG nova.virt.hardware [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 657.024651] env[62965]: DEBUG nova.virt.hardware [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 657.024922] env[62965]: DEBUG nova.virt.hardware [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 657.025772] env[62965]: DEBUG nova.virt.hardware [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 657.027205] env[62965]: DEBUG nova.virt.hardware [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 657.027205] env[62965]: DEBUG nova.virt.hardware [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 657.027205] env[62965]: DEBUG nova.virt.hardware [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 657.028974] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7786533-2920-4179-9384-516cd7b8189e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.040118] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13e222e8-2320-45e3-b21e-4c7219390b1e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.162314] env[62965]: ERROR nova.compute.manager [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 763d2a98-a632-43dd-8f7e-f3c96720096e, please check neutron logs for more information. [ 657.162314] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 657.162314] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 657.162314] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 657.162314] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 657.162314] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 657.162314] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 657.162314] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 657.162314] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.162314] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 657.162314] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.162314] env[62965]: ERROR nova.compute.manager raise self.value [ 657.162314] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 657.162314] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 657.162314] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.162314] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 657.162777] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.162777] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 657.162777] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 763d2a98-a632-43dd-8f7e-f3c96720096e, please check neutron logs for more information. [ 657.162777] env[62965]: ERROR nova.compute.manager [ 657.162777] env[62965]: Traceback (most recent call last): [ 657.162777] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 657.162777] env[62965]: listener.cb(fileno) [ 657.162777] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 657.162777] env[62965]: result = function(*args, **kwargs) [ 657.162777] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 657.162777] env[62965]: return func(*args, **kwargs) [ 657.162777] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 657.162777] env[62965]: raise e [ 657.162777] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 657.162777] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 657.162777] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 657.162777] env[62965]: created_port_ids = self._update_ports_for_instance( [ 657.162777] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 657.162777] env[62965]: with excutils.save_and_reraise_exception(): [ 657.162777] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.162777] env[62965]: self.force_reraise() [ 657.162777] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.162777] env[62965]: raise self.value [ 657.162777] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 657.162777] env[62965]: updated_port = self._update_port( [ 657.162777] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.162777] env[62965]: _ensure_no_port_binding_failure(port) [ 657.162777] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.162777] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 657.163485] env[62965]: nova.exception.PortBindingFailed: Binding failed for port 763d2a98-a632-43dd-8f7e-f3c96720096e, please check neutron logs for more information. [ 657.163485] env[62965]: Removing descriptor: 19 [ 657.163485] env[62965]: ERROR nova.compute.manager [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 763d2a98-a632-43dd-8f7e-f3c96720096e, please check neutron logs for more information. [ 657.163485] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Traceback (most recent call last): [ 657.163485] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 657.163485] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] yield resources [ 657.163485] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 657.163485] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] self.driver.spawn(context, instance, image_meta, [ 657.163485] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 657.163485] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] self._vmops.spawn(context, instance, image_meta, injected_files, [ 657.163485] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 657.163485] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] vm_ref = self.build_virtual_machine(instance, [ 657.164041] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 657.164041] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] vif_infos = vmwarevif.get_vif_info(self._session, [ 657.164041] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 657.164041] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] for vif in network_info: [ 657.164041] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 657.164041] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] return self._sync_wrapper(fn, *args, **kwargs) [ 657.164041] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 657.164041] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] self.wait() [ 657.164041] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 657.164041] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] self[:] = self._gt.wait() [ 657.164041] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 657.164041] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] return self._exit_event.wait() [ 657.164041] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 657.164410] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] result = hub.switch() [ 657.164410] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 657.164410] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] return self.greenlet.switch() [ 657.164410] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 657.164410] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] result = function(*args, **kwargs) [ 657.164410] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 657.164410] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] return func(*args, **kwargs) [ 657.164410] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 657.164410] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] raise e [ 657.164410] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 657.164410] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] nwinfo = self.network_api.allocate_for_instance( [ 657.164410] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 657.164410] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] created_port_ids = self._update_ports_for_instance( [ 657.164741] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 657.164741] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] with excutils.save_and_reraise_exception(): [ 657.164741] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.164741] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] self.force_reraise() [ 657.164741] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.164741] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] raise self.value [ 657.164741] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 657.164741] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] updated_port = self._update_port( [ 657.164741] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.164741] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] _ensure_no_port_binding_failure(port) [ 657.164741] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.164741] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] raise exception.PortBindingFailed(port_id=port['id']) [ 657.165142] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] nova.exception.PortBindingFailed: Binding failed for port 763d2a98-a632-43dd-8f7e-f3c96720096e, please check neutron logs for more information. [ 657.165142] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] [ 657.165142] env[62965]: INFO nova.compute.manager [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Terminating instance [ 657.181606] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.470019] env[62965]: DEBUG nova.compute.manager [req-6505199d-7501-4305-bff3-c902bd374a3e req-f2f2413e-77dd-4ff5-8622-4612fe901b91 service nova] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Received event network-changed-763d2a98-a632-43dd-8f7e-f3c96720096e {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 657.470019] env[62965]: DEBUG nova.compute.manager [req-6505199d-7501-4305-bff3-c902bd374a3e req-f2f2413e-77dd-4ff5-8622-4612fe901b91 service nova] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Refreshing instance network info cache due to event network-changed-763d2a98-a632-43dd-8f7e-f3c96720096e. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 657.470019] env[62965]: DEBUG oslo_concurrency.lockutils [req-6505199d-7501-4305-bff3-c902bd374a3e req-f2f2413e-77dd-4ff5-8622-4612fe901b91 service nova] Acquiring lock "refresh_cache-8ac4264d-7d75-4c38-b8da-e3d90ffe9363" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 657.470019] env[62965]: DEBUG oslo_concurrency.lockutils [req-6505199d-7501-4305-bff3-c902bd374a3e req-f2f2413e-77dd-4ff5-8622-4612fe901b91 service nova] Acquired lock "refresh_cache-8ac4264d-7d75-4c38-b8da-e3d90ffe9363" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.470019] env[62965]: DEBUG nova.network.neutron [req-6505199d-7501-4305-bff3-c902bd374a3e req-f2f2413e-77dd-4ff5-8622-4612fe901b91 service nova] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Refreshing network info cache for port 763d2a98-a632-43dd-8f7e-f3c96720096e {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 657.508422] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.557s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 657.509220] env[62965]: DEBUG nova.compute.manager [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 657.514388] env[62965]: DEBUG oslo_concurrency.lockutils [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.177s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.518180] env[62965]: INFO nova.compute.claims [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 657.672316] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] Acquiring lock "refresh_cache-8ac4264d-7d75-4c38-b8da-e3d90ffe9363" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 658.004288] env[62965]: DEBUG nova.network.neutron [req-6505199d-7501-4305-bff3-c902bd374a3e req-f2f2413e-77dd-4ff5-8622-4612fe901b91 service nova] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 658.024321] env[62965]: DEBUG nova.compute.utils [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 658.024840] env[62965]: DEBUG nova.compute.manager [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 658.025059] env[62965]: DEBUG nova.network.neutron [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 658.101333] env[62965]: DEBUG nova.policy [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7dca5bc95b22405b8f04af2808277009', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '346df53fb497435b9e0c1157c502e8e9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 658.219696] env[62965]: DEBUG nova.network.neutron [req-6505199d-7501-4305-bff3-c902bd374a3e req-f2f2413e-77dd-4ff5-8622-4612fe901b91 service nova] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.526179] env[62965]: DEBUG nova.compute.manager [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 658.680943] env[62965]: DEBUG nova.network.neutron [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Successfully created port: 1da29949-9732-45b3-a088-7b379ed73ad5 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 658.725402] env[62965]: DEBUG oslo_concurrency.lockutils [req-6505199d-7501-4305-bff3-c902bd374a3e req-f2f2413e-77dd-4ff5-8622-4612fe901b91 service nova] Releasing lock "refresh_cache-8ac4264d-7d75-4c38-b8da-e3d90ffe9363" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 658.725649] env[62965]: DEBUG nova.compute.manager [req-6505199d-7501-4305-bff3-c902bd374a3e req-f2f2413e-77dd-4ff5-8622-4612fe901b91 service nova] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Received event network-vif-deleted-763d2a98-a632-43dd-8f7e-f3c96720096e {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 658.726979] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] Acquired lock "refresh_cache-8ac4264d-7d75-4c38-b8da-e3d90ffe9363" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 658.727176] env[62965]: DEBUG nova.network.neutron [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 658.980495] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bf18e0a-dc99-4ade-a10b-2abea77b2bc7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.988595] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fdf04be-a7b1-4d1e-9bbb-0d8594320371 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.025736] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19aa901c-362f-4238-9457-8c48e264a56d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.034055] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63ddb6cb-e824-4c3b-8994-fdbb0a388dc2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.052815] env[62965]: DEBUG nova.compute.provider_tree [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 659.265873] env[62965]: DEBUG nova.network.neutron [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 659.435140] env[62965]: DEBUG nova.network.neutron [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.544616] env[62965]: DEBUG nova.compute.manager [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 659.557625] env[62965]: DEBUG nova.scheduler.client.report [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 659.593013] env[62965]: DEBUG nova.virt.hardware [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 659.593013] env[62965]: DEBUG nova.virt.hardware [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 659.593013] env[62965]: DEBUG nova.virt.hardware [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 659.593213] env[62965]: DEBUG nova.virt.hardware [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 659.593213] env[62965]: DEBUG nova.virt.hardware [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 659.593213] env[62965]: DEBUG nova.virt.hardware [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 659.593213] env[62965]: DEBUG nova.virt.hardware [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 659.593439] env[62965]: DEBUG nova.virt.hardware [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 659.593751] env[62965]: DEBUG nova.virt.hardware [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 659.594056] env[62965]: DEBUG nova.virt.hardware [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 659.594454] env[62965]: DEBUG nova.virt.hardware [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 659.599558] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bd142a5-4134-4c30-9128-097926655081 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.607500] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3ddbbaa-3c87-47e1-8ac9-3ada5991ee10 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.945221] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] Releasing lock "refresh_cache-8ac4264d-7d75-4c38-b8da-e3d90ffe9363" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 659.945221] env[62965]: DEBUG nova.compute.manager [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 659.945490] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 659.946424] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-94d5cae7-f65f-4a23-a94b-e709129fe8b8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.959691] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b6133c3-85c7-434a-8963-7d582a6ecb74 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.985379] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8ac4264d-7d75-4c38-b8da-e3d90ffe9363 could not be found. [ 659.985675] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 659.985915] env[62965]: INFO nova.compute.manager [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Took 0.04 seconds to destroy the instance on the hypervisor. [ 659.986241] env[62965]: DEBUG oslo.service.loopingcall [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 659.986682] env[62965]: DEBUG nova.compute.manager [-] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 659.986850] env[62965]: DEBUG nova.network.neutron [-] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 660.020357] env[62965]: DEBUG nova.network.neutron [-] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 660.067134] env[62965]: DEBUG oslo_concurrency.lockutils [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.550s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.067134] env[62965]: DEBUG nova.compute.manager [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 660.068371] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 17.180s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.068548] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.068704] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62965) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 660.068970] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.784s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.070873] env[62965]: INFO nova.compute.claims [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 660.077070] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05e4fc54-e1af-4a51-86e5-2936daee48c8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.085916] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63ac2943-e8b4-41c2-bb62-48e6c3d644ce {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.102557] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eeed637-0130-4567-af0b-f96ab48c7318 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.113714] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a71eced-1c50-4f1e-90e0-fad5324d2361 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.144873] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181372MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62965) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 660.145053] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 660.373650] env[62965]: DEBUG nova.compute.manager [req-2cb8e8f1-e0f4-4645-b5bf-0b36c119a307 req-de893526-35fe-4811-9c10-dadd92a7e8a0 service nova] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Received event network-changed-1da29949-9732-45b3-a088-7b379ed73ad5 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 660.373650] env[62965]: DEBUG nova.compute.manager [req-2cb8e8f1-e0f4-4645-b5bf-0b36c119a307 req-de893526-35fe-4811-9c10-dadd92a7e8a0 service nova] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Refreshing instance network info cache due to event network-changed-1da29949-9732-45b3-a088-7b379ed73ad5. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 660.373650] env[62965]: DEBUG oslo_concurrency.lockutils [req-2cb8e8f1-e0f4-4645-b5bf-0b36c119a307 req-de893526-35fe-4811-9c10-dadd92a7e8a0 service nova] Acquiring lock "refresh_cache-971f72f8-2686-4ced-a2a1-b7ab8a1ea024" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.373650] env[62965]: DEBUG oslo_concurrency.lockutils [req-2cb8e8f1-e0f4-4645-b5bf-0b36c119a307 req-de893526-35fe-4811-9c10-dadd92a7e8a0 service nova] Acquired lock "refresh_cache-971f72f8-2686-4ced-a2a1-b7ab8a1ea024" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.373650] env[62965]: DEBUG nova.network.neutron [req-2cb8e8f1-e0f4-4645-b5bf-0b36c119a307 req-de893526-35fe-4811-9c10-dadd92a7e8a0 service nova] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Refreshing network info cache for port 1da29949-9732-45b3-a088-7b379ed73ad5 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 660.521572] env[62965]: DEBUG nova.network.neutron [-] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.578209] env[62965]: DEBUG nova.compute.utils [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 660.580303] env[62965]: DEBUG nova.compute.manager [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 660.583968] env[62965]: DEBUG nova.network.neutron [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 660.621086] env[62965]: ERROR nova.compute.manager [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1da29949-9732-45b3-a088-7b379ed73ad5, please check neutron logs for more information. [ 660.621086] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 660.621086] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 660.621086] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 660.621086] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 660.621086] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 660.621086] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 660.621086] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 660.621086] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.621086] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 660.621086] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.621086] env[62965]: ERROR nova.compute.manager raise self.value [ 660.621086] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 660.621086] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 660.621086] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.621086] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 660.621589] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.621589] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 660.621589] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1da29949-9732-45b3-a088-7b379ed73ad5, please check neutron logs for more information. [ 660.621589] env[62965]: ERROR nova.compute.manager [ 660.621589] env[62965]: Traceback (most recent call last): [ 660.621589] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 660.621589] env[62965]: listener.cb(fileno) [ 660.621589] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 660.621589] env[62965]: result = function(*args, **kwargs) [ 660.621589] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 660.621589] env[62965]: return func(*args, **kwargs) [ 660.621589] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 660.621589] env[62965]: raise e [ 660.621589] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 660.621589] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 660.621589] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 660.621589] env[62965]: created_port_ids = self._update_ports_for_instance( [ 660.621589] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 660.621589] env[62965]: with excutils.save_and_reraise_exception(): [ 660.621589] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.621589] env[62965]: self.force_reraise() [ 660.621589] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.621589] env[62965]: raise self.value [ 660.621589] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 660.621589] env[62965]: updated_port = self._update_port( [ 660.621589] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.621589] env[62965]: _ensure_no_port_binding_failure(port) [ 660.621589] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.621589] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 660.622358] env[62965]: nova.exception.PortBindingFailed: Binding failed for port 1da29949-9732-45b3-a088-7b379ed73ad5, please check neutron logs for more information. [ 660.622358] env[62965]: Removing descriptor: 17 [ 660.622358] env[62965]: ERROR nova.compute.manager [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1da29949-9732-45b3-a088-7b379ed73ad5, please check neutron logs for more information. [ 660.622358] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Traceback (most recent call last): [ 660.622358] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 660.622358] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] yield resources [ 660.622358] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 660.622358] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] self.driver.spawn(context, instance, image_meta, [ 660.622358] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 660.622358] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] self._vmops.spawn(context, instance, image_meta, injected_files, [ 660.622358] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 660.622358] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] vm_ref = self.build_virtual_machine(instance, [ 660.622689] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 660.622689] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] vif_infos = vmwarevif.get_vif_info(self._session, [ 660.622689] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 660.622689] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] for vif in network_info: [ 660.622689] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 660.622689] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] return self._sync_wrapper(fn, *args, **kwargs) [ 660.622689] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 660.622689] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] self.wait() [ 660.622689] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 660.622689] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] self[:] = self._gt.wait() [ 660.622689] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 660.622689] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] return self._exit_event.wait() [ 660.622689] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 660.623041] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] result = hub.switch() [ 660.623041] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 660.623041] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] return self.greenlet.switch() [ 660.623041] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 660.623041] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] result = function(*args, **kwargs) [ 660.623041] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 660.623041] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] return func(*args, **kwargs) [ 660.623041] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 660.623041] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] raise e [ 660.623041] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 660.623041] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] nwinfo = self.network_api.allocate_for_instance( [ 660.623041] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 660.623041] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] created_port_ids = self._update_ports_for_instance( [ 660.623382] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 660.623382] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] with excutils.save_and_reraise_exception(): [ 660.623382] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 660.623382] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] self.force_reraise() [ 660.623382] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 660.623382] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] raise self.value [ 660.623382] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 660.623382] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] updated_port = self._update_port( [ 660.623382] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 660.623382] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] _ensure_no_port_binding_failure(port) [ 660.623382] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 660.623382] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] raise exception.PortBindingFailed(port_id=port['id']) [ 660.623698] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] nova.exception.PortBindingFailed: Binding failed for port 1da29949-9732-45b3-a088-7b379ed73ad5, please check neutron logs for more information. [ 660.623698] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] [ 660.623698] env[62965]: INFO nova.compute.manager [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Terminating instance [ 660.641294] env[62965]: DEBUG nova.policy [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5d1bf4c1dbfa46dc9ed97a136cec314f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '295fdf21de9b4c999b79781a8aa3726b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 660.905037] env[62965]: DEBUG nova.network.neutron [req-2cb8e8f1-e0f4-4645-b5bf-0b36c119a307 req-de893526-35fe-4811-9c10-dadd92a7e8a0 service nova] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 661.023727] env[62965]: INFO nova.compute.manager [-] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Took 1.04 seconds to deallocate network for instance. [ 661.026381] env[62965]: DEBUG nova.compute.claims [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 661.026561] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 661.087122] env[62965]: DEBUG nova.compute.manager [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 661.112354] env[62965]: DEBUG nova.network.neutron [req-2cb8e8f1-e0f4-4645-b5bf-0b36c119a307 req-de893526-35fe-4811-9c10-dadd92a7e8a0 service nova] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.120870] env[62965]: DEBUG nova.network.neutron [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Successfully created port: 09855567-2998-47e2-9b3c-5e2f7b42f71f {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 661.131897] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Acquiring lock "refresh_cache-971f72f8-2686-4ced-a2a1-b7ab8a1ea024" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 661.618847] env[62965]: DEBUG oslo_concurrency.lockutils [req-2cb8e8f1-e0f4-4645-b5bf-0b36c119a307 req-de893526-35fe-4811-9c10-dadd92a7e8a0 service nova] Releasing lock "refresh_cache-971f72f8-2686-4ced-a2a1-b7ab8a1ea024" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.619759] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Acquired lock "refresh_cache-971f72f8-2686-4ced-a2a1-b7ab8a1ea024" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.620758] env[62965]: DEBUG nova.network.neutron [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 661.735522] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2c5a8e5-668d-48ee-8b91-b763b47ec6ef {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.743507] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7796bbee-e640-460f-901b-20bc754050ce {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.773288] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa355b02-8fc7-4909-8b33-58e5710b01a1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.780539] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8afb97ff-c2ed-475b-9b82-c9ffd846c670 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.795416] env[62965]: DEBUG nova.compute.provider_tree [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 662.107300] env[62965]: DEBUG nova.compute.manager [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 662.145077] env[62965]: DEBUG nova.virt.hardware [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 662.145077] env[62965]: DEBUG nova.virt.hardware [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 662.145077] env[62965]: DEBUG nova.virt.hardware [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 662.145394] env[62965]: DEBUG nova.virt.hardware [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 662.145394] env[62965]: DEBUG nova.virt.hardware [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 662.145394] env[62965]: DEBUG nova.virt.hardware [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 662.145394] env[62965]: DEBUG nova.virt.hardware [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 662.148663] env[62965]: DEBUG nova.virt.hardware [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 662.150230] env[62965]: DEBUG nova.virt.hardware [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 662.150230] env[62965]: DEBUG nova.virt.hardware [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 662.150230] env[62965]: DEBUG nova.virt.hardware [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 662.150230] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10ce4d9b-4400-4a3d-9ad7-dcf8d7245e88 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.154665] env[62965]: DEBUG nova.network.neutron [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 662.161810] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50856577-574d-4f78-8bef-838aa7a62b47 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.265021] env[62965]: DEBUG nova.network.neutron [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.301017] env[62965]: DEBUG nova.scheduler.client.report [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 662.446173] env[62965]: DEBUG nova.compute.manager [req-3599bf06-672e-44b3-8018-aea8cc9ad668 req-e744dc24-5f38-4b51-9b37-31534543c476 service nova] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Received event network-vif-deleted-1da29949-9732-45b3-a088-7b379ed73ad5 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 662.766134] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Releasing lock "refresh_cache-971f72f8-2686-4ced-a2a1-b7ab8a1ea024" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 662.766566] env[62965]: DEBUG nova.compute.manager [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 662.766758] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 662.767080] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ba8b5e61-9b2e-4b8b-ada7-c08b9b616770 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.776390] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1b21d39-f6b8-4127-9efe-364f1b3a1033 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.800294] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 971f72f8-2686-4ced-a2a1-b7ab8a1ea024 could not be found. [ 662.800294] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 662.800294] env[62965]: INFO nova.compute.manager [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Took 0.03 seconds to destroy the instance on the hypervisor. [ 662.800294] env[62965]: DEBUG oslo.service.loopingcall [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 662.800631] env[62965]: DEBUG nova.compute.manager [-] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 662.800755] env[62965]: DEBUG nova.network.neutron [-] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 662.806276] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.737s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 662.806792] env[62965]: DEBUG nova.compute.manager [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 662.809837] env[62965]: DEBUG oslo_concurrency.lockutils [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.517s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 662.810806] env[62965]: INFO nova.compute.claims [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 662.829978] env[62965]: DEBUG nova.network.neutron [-] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 663.237227] env[62965]: DEBUG nova.compute.manager [req-66eecc02-59cc-46dd-a463-ceb672521863 req-dbd3398e-09f6-4d1e-8191-7f375801091d service nova] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Received event network-changed-09855567-2998-47e2-9b3c-5e2f7b42f71f {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 663.237227] env[62965]: DEBUG nova.compute.manager [req-66eecc02-59cc-46dd-a463-ceb672521863 req-dbd3398e-09f6-4d1e-8191-7f375801091d service nova] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Refreshing instance network info cache due to event network-changed-09855567-2998-47e2-9b3c-5e2f7b42f71f. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 663.237227] env[62965]: DEBUG oslo_concurrency.lockutils [req-66eecc02-59cc-46dd-a463-ceb672521863 req-dbd3398e-09f6-4d1e-8191-7f375801091d service nova] Acquiring lock "refresh_cache-eaf9ca94-4f45-403e-bb06-e9a9adf58f38" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 663.237227] env[62965]: DEBUG oslo_concurrency.lockutils [req-66eecc02-59cc-46dd-a463-ceb672521863 req-dbd3398e-09f6-4d1e-8191-7f375801091d service nova] Acquired lock "refresh_cache-eaf9ca94-4f45-403e-bb06-e9a9adf58f38" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.237227] env[62965]: DEBUG nova.network.neutron [req-66eecc02-59cc-46dd-a463-ceb672521863 req-dbd3398e-09f6-4d1e-8191-7f375801091d service nova] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Refreshing network info cache for port 09855567-2998-47e2-9b3c-5e2f7b42f71f {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 663.314989] env[62965]: DEBUG nova.compute.utils [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 663.319333] env[62965]: DEBUG nova.compute.manager [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 663.320018] env[62965]: DEBUG nova.network.neutron [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 663.332966] env[62965]: DEBUG nova.network.neutron [-] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.430163] env[62965]: DEBUG nova.policy [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7dca5bc95b22405b8f04af2808277009', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '346df53fb497435b9e0c1157c502e8e9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 663.547692] env[62965]: ERROR nova.compute.manager [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 09855567-2998-47e2-9b3c-5e2f7b42f71f, please check neutron logs for more information. [ 663.547692] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 663.547692] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 663.547692] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 663.547692] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 663.547692] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 663.547692] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 663.547692] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 663.547692] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.547692] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 663.547692] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.547692] env[62965]: ERROR nova.compute.manager raise self.value [ 663.547692] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 663.547692] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 663.547692] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.547692] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 663.548209] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.548209] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 663.548209] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 09855567-2998-47e2-9b3c-5e2f7b42f71f, please check neutron logs for more information. [ 663.548209] env[62965]: ERROR nova.compute.manager [ 663.548209] env[62965]: Traceback (most recent call last): [ 663.548209] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 663.548209] env[62965]: listener.cb(fileno) [ 663.548209] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 663.548209] env[62965]: result = function(*args, **kwargs) [ 663.548209] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 663.548209] env[62965]: return func(*args, **kwargs) [ 663.548209] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 663.548209] env[62965]: raise e [ 663.548209] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 663.548209] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 663.548209] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 663.548209] env[62965]: created_port_ids = self._update_ports_for_instance( [ 663.548209] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 663.548209] env[62965]: with excutils.save_and_reraise_exception(): [ 663.548209] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.548209] env[62965]: self.force_reraise() [ 663.548209] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.548209] env[62965]: raise self.value [ 663.548209] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 663.548209] env[62965]: updated_port = self._update_port( [ 663.548209] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.548209] env[62965]: _ensure_no_port_binding_failure(port) [ 663.548209] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.548209] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 663.548937] env[62965]: nova.exception.PortBindingFailed: Binding failed for port 09855567-2998-47e2-9b3c-5e2f7b42f71f, please check neutron logs for more information. [ 663.548937] env[62965]: Removing descriptor: 19 [ 663.552104] env[62965]: ERROR nova.compute.manager [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 09855567-2998-47e2-9b3c-5e2f7b42f71f, please check neutron logs for more information. [ 663.552104] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Traceback (most recent call last): [ 663.552104] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 663.552104] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] yield resources [ 663.552104] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 663.552104] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] self.driver.spawn(context, instance, image_meta, [ 663.552104] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 663.552104] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] self._vmops.spawn(context, instance, image_meta, injected_files, [ 663.552104] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 663.552104] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] vm_ref = self.build_virtual_machine(instance, [ 663.552104] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 663.552466] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] vif_infos = vmwarevif.get_vif_info(self._session, [ 663.552466] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 663.552466] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] for vif in network_info: [ 663.552466] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 663.552466] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] return self._sync_wrapper(fn, *args, **kwargs) [ 663.552466] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 663.552466] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] self.wait() [ 663.552466] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 663.552466] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] self[:] = self._gt.wait() [ 663.552466] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 663.552466] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] return self._exit_event.wait() [ 663.552466] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 663.552466] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] result = hub.switch() [ 663.552815] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 663.552815] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] return self.greenlet.switch() [ 663.552815] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 663.552815] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] result = function(*args, **kwargs) [ 663.552815] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 663.552815] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] return func(*args, **kwargs) [ 663.552815] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 663.552815] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] raise e [ 663.552815] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 663.552815] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] nwinfo = self.network_api.allocate_for_instance( [ 663.552815] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 663.552815] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] created_port_ids = self._update_ports_for_instance( [ 663.552815] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 663.553207] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] with excutils.save_and_reraise_exception(): [ 663.553207] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.553207] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] self.force_reraise() [ 663.553207] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.553207] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] raise self.value [ 663.553207] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 663.553207] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] updated_port = self._update_port( [ 663.553207] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.553207] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] _ensure_no_port_binding_failure(port) [ 663.553207] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.553207] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] raise exception.PortBindingFailed(port_id=port['id']) [ 663.553207] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] nova.exception.PortBindingFailed: Binding failed for port 09855567-2998-47e2-9b3c-5e2f7b42f71f, please check neutron logs for more information. [ 663.553207] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] [ 663.553672] env[62965]: INFO nova.compute.manager [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Terminating instance [ 663.827187] env[62965]: DEBUG nova.compute.manager [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 663.836346] env[62965]: INFO nova.compute.manager [-] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Took 1.03 seconds to deallocate network for instance. [ 663.840367] env[62965]: DEBUG nova.compute.claims [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 663.840367] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 663.911805] env[62965]: DEBUG nova.network.neutron [req-66eecc02-59cc-46dd-a463-ceb672521863 req-dbd3398e-09f6-4d1e-8191-7f375801091d service nova] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 664.048388] env[62965]: DEBUG nova.network.neutron [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Successfully created port: 76d8df8a-8e64-4dd2-b6bd-f1726c488d7f {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 664.057301] env[62965]: DEBUG oslo_concurrency.lockutils [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] Acquiring lock "refresh_cache-eaf9ca94-4f45-403e-bb06-e9a9adf58f38" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 664.071038] env[62965]: DEBUG nova.network.neutron [req-66eecc02-59cc-46dd-a463-ceb672521863 req-dbd3398e-09f6-4d1e-8191-7f375801091d service nova] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.291770] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3001f68-2040-43c5-af7f-3b369b3e2224 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.302714] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0624d931-081e-4228-a5df-7897a6daf644 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.339585] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65aff1ae-a9ac-452d-8cfc-c25b0dc6cbd5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.348897] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-445a720d-7514-48ca-bb6a-f38d443fb7bb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.362442] env[62965]: DEBUG nova.compute.provider_tree [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 664.573721] env[62965]: DEBUG oslo_concurrency.lockutils [req-66eecc02-59cc-46dd-a463-ceb672521863 req-dbd3398e-09f6-4d1e-8191-7f375801091d service nova] Releasing lock "refresh_cache-eaf9ca94-4f45-403e-bb06-e9a9adf58f38" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 664.574252] env[62965]: DEBUG oslo_concurrency.lockutils [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] Acquired lock "refresh_cache-eaf9ca94-4f45-403e-bb06-e9a9adf58f38" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.574728] env[62965]: DEBUG nova.network.neutron [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 664.843868] env[62965]: DEBUG nova.compute.manager [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 664.866580] env[62965]: DEBUG nova.scheduler.client.report [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 664.883258] env[62965]: DEBUG nova.virt.hardware [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:24Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 664.883482] env[62965]: DEBUG nova.virt.hardware [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 664.883636] env[62965]: DEBUG nova.virt.hardware [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 664.883807] env[62965]: DEBUG nova.virt.hardware [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 664.883947] env[62965]: DEBUG nova.virt.hardware [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 664.885670] env[62965]: DEBUG nova.virt.hardware [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 664.886890] env[62965]: DEBUG nova.virt.hardware [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 664.886890] env[62965]: DEBUG nova.virt.hardware [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 664.886890] env[62965]: DEBUG nova.virt.hardware [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 664.886890] env[62965]: DEBUG nova.virt.hardware [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 664.886890] env[62965]: DEBUG nova.virt.hardware [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 664.888264] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdce6a08-fa43-4c9d-b9f2-8b9bda00da3e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.902517] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7f594bd-e9d6-4ad9-b31e-d7c073070db2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.102255] env[62965]: DEBUG nova.network.neutron [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 665.214400] env[62965]: DEBUG nova.network.neutron [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.376096] env[62965]: DEBUG oslo_concurrency.lockutils [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.567s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 665.376633] env[62965]: DEBUG nova.compute.manager [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 665.384018] env[62965]: DEBUG oslo_concurrency.lockutils [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.064s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 665.603864] env[62965]: DEBUG nova.compute.manager [req-6292326b-63fb-434f-b7ca-58d8964fa96e req-16bd59e2-c0ab-472e-b586-03bfa8f456ad service nova] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Received event network-vif-deleted-09855567-2998-47e2-9b3c-5e2f7b42f71f {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 665.718360] env[62965]: DEBUG oslo_concurrency.lockutils [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] Releasing lock "refresh_cache-eaf9ca94-4f45-403e-bb06-e9a9adf58f38" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 665.718821] env[62965]: DEBUG nova.compute.manager [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 665.719661] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 665.720052] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e3bc3d9f-b7d4-4fd0-8eb4-f82c97ceccc5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.731919] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb9de720-25e5-425c-9f92-47e8ba47d316 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.757936] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance eaf9ca94-4f45-403e-bb06-e9a9adf58f38 could not be found. [ 665.757936] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 665.757936] env[62965]: INFO nova.compute.manager [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Took 0.04 seconds to destroy the instance on the hypervisor. [ 665.758056] env[62965]: DEBUG oslo.service.loopingcall [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 665.758383] env[62965]: DEBUG nova.compute.manager [-] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 665.758483] env[62965]: DEBUG nova.network.neutron [-] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 665.798341] env[62965]: DEBUG nova.network.neutron [-] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 665.884553] env[62965]: DEBUG nova.compute.utils [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 665.886343] env[62965]: DEBUG nova.compute.manager [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 665.886512] env[62965]: DEBUG nova.network.neutron [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 665.894291] env[62965]: DEBUG nova.compute.manager [req-06aff75f-3744-488b-b3c4-93ca07625cb0 req-31e78353-894e-42ac-ba6a-17f3bd7edf74 service nova] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Received event network-changed-76d8df8a-8e64-4dd2-b6bd-f1726c488d7f {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 665.894916] env[62965]: DEBUG nova.compute.manager [req-06aff75f-3744-488b-b3c4-93ca07625cb0 req-31e78353-894e-42ac-ba6a-17f3bd7edf74 service nova] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Refreshing instance network info cache due to event network-changed-76d8df8a-8e64-4dd2-b6bd-f1726c488d7f. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 665.894916] env[62965]: DEBUG oslo_concurrency.lockutils [req-06aff75f-3744-488b-b3c4-93ca07625cb0 req-31e78353-894e-42ac-ba6a-17f3bd7edf74 service nova] Acquiring lock "refresh_cache-79a4d3c8-ead6-4893-b0a7-89a8ff435e54" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 665.894916] env[62965]: DEBUG oslo_concurrency.lockutils [req-06aff75f-3744-488b-b3c4-93ca07625cb0 req-31e78353-894e-42ac-ba6a-17f3bd7edf74 service nova] Acquired lock "refresh_cache-79a4d3c8-ead6-4893-b0a7-89a8ff435e54" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.894916] env[62965]: DEBUG nova.network.neutron [req-06aff75f-3744-488b-b3c4-93ca07625cb0 req-31e78353-894e-42ac-ba6a-17f3bd7edf74 service nova] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Refreshing network info cache for port 76d8df8a-8e64-4dd2-b6bd-f1726c488d7f {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 665.950113] env[62965]: DEBUG nova.policy [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '96279b4d04cb46949ccc343b4a6198da', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '63609764674145849796434c49d2299c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 666.106424] env[62965]: ERROR nova.compute.manager [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 76d8df8a-8e64-4dd2-b6bd-f1726c488d7f, please check neutron logs for more information. [ 666.106424] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 666.106424] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 666.106424] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 666.106424] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 666.106424] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 666.106424] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 666.106424] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 666.106424] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 666.106424] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 666.106424] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 666.106424] env[62965]: ERROR nova.compute.manager raise self.value [ 666.106424] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 666.106424] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 666.106424] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 666.106424] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 666.106912] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 666.106912] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 666.106912] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 76d8df8a-8e64-4dd2-b6bd-f1726c488d7f, please check neutron logs for more information. [ 666.106912] env[62965]: ERROR nova.compute.manager [ 666.107738] env[62965]: Traceback (most recent call last): [ 666.108097] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 666.108097] env[62965]: listener.cb(fileno) [ 666.108097] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 666.108097] env[62965]: result = function(*args, **kwargs) [ 666.108097] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 666.108097] env[62965]: return func(*args, **kwargs) [ 666.108097] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 666.108097] env[62965]: raise e [ 666.108097] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 666.108097] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 666.108097] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 666.108097] env[62965]: created_port_ids = self._update_ports_for_instance( [ 666.108097] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 666.108097] env[62965]: with excutils.save_and_reraise_exception(): [ 666.108097] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 666.108097] env[62965]: self.force_reraise() [ 666.108097] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 666.108097] env[62965]: raise self.value [ 666.108097] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 666.108097] env[62965]: updated_port = self._update_port( [ 666.108097] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 666.108097] env[62965]: _ensure_no_port_binding_failure(port) [ 666.108097] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 666.108097] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 666.108097] env[62965]: nova.exception.PortBindingFailed: Binding failed for port 76d8df8a-8e64-4dd2-b6bd-f1726c488d7f, please check neutron logs for more information. [ 666.108097] env[62965]: Removing descriptor: 17 [ 666.111845] env[62965]: ERROR nova.compute.manager [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 76d8df8a-8e64-4dd2-b6bd-f1726c488d7f, please check neutron logs for more information. [ 666.111845] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Traceback (most recent call last): [ 666.111845] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 666.111845] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] yield resources [ 666.111845] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 666.111845] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] self.driver.spawn(context, instance, image_meta, [ 666.111845] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 666.111845] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] self._vmops.spawn(context, instance, image_meta, injected_files, [ 666.111845] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 666.111845] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] vm_ref = self.build_virtual_machine(instance, [ 666.111845] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 666.113058] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] vif_infos = vmwarevif.get_vif_info(self._session, [ 666.113058] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 666.113058] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] for vif in network_info: [ 666.113058] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 666.113058] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] return self._sync_wrapper(fn, *args, **kwargs) [ 666.113058] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 666.113058] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] self.wait() [ 666.113058] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 666.113058] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] self[:] = self._gt.wait() [ 666.113058] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 666.113058] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] return self._exit_event.wait() [ 666.113058] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 666.113058] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] result = hub.switch() [ 666.113441] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 666.113441] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] return self.greenlet.switch() [ 666.113441] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 666.113441] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] result = function(*args, **kwargs) [ 666.113441] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 666.113441] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] return func(*args, **kwargs) [ 666.113441] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 666.113441] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] raise e [ 666.113441] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 666.113441] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] nwinfo = self.network_api.allocate_for_instance( [ 666.113441] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 666.113441] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] created_port_ids = self._update_ports_for_instance( [ 666.113441] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 666.113800] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] with excutils.save_and_reraise_exception(): [ 666.113800] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 666.113800] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] self.force_reraise() [ 666.113800] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 666.113800] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] raise self.value [ 666.113800] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 666.113800] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] updated_port = self._update_port( [ 666.113800] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 666.113800] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] _ensure_no_port_binding_failure(port) [ 666.113800] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 666.113800] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] raise exception.PortBindingFailed(port_id=port['id']) [ 666.113800] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] nova.exception.PortBindingFailed: Binding failed for port 76d8df8a-8e64-4dd2-b6bd-f1726c488d7f, please check neutron logs for more information. [ 666.113800] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] [ 666.114830] env[62965]: INFO nova.compute.manager [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Terminating instance [ 666.303202] env[62965]: DEBUG nova.network.neutron [-] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.351034] env[62965]: DEBUG nova.network.neutron [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Successfully created port: 85c8faec-56d2-4fad-86ff-cc4710a93278 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 666.393344] env[62965]: DEBUG nova.compute.manager [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 666.395555] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bb476fa-051d-4c7b-ad76-521c51bf3e94 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.407579] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a82b2077-dca9-409c-9d6b-73e0c34f66e2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.444838] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "dcae9cae-d4f5-4b34-b392-00b5ef9c81bf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 666.445154] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "dcae9cae-d4f5-4b34-b392-00b5ef9c81bf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 666.446220] env[62965]: DEBUG nova.network.neutron [req-06aff75f-3744-488b-b3c4-93ca07625cb0 req-31e78353-894e-42ac-ba6a-17f3bd7edf74 service nova] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 666.448567] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa7a6301-595e-456f-a4e9-c363227ebb89 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.457640] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-778d9cb2-7d6e-413a-ab68-1477984d1122 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.472189] env[62965]: DEBUG nova.compute.provider_tree [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 666.554124] env[62965]: DEBUG nova.network.neutron [req-06aff75f-3744-488b-b3c4-93ca07625cb0 req-31e78353-894e-42ac-ba6a-17f3bd7edf74 service nova] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.618329] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Acquiring lock "refresh_cache-79a4d3c8-ead6-4893-b0a7-89a8ff435e54" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 666.807371] env[62965]: INFO nova.compute.manager [-] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Took 1.05 seconds to deallocate network for instance. [ 666.811713] env[62965]: DEBUG nova.compute.claims [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 666.811713] env[62965]: DEBUG oslo_concurrency.lockutils [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 666.979916] env[62965]: DEBUG nova.scheduler.client.report [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 667.060216] env[62965]: DEBUG oslo_concurrency.lockutils [req-06aff75f-3744-488b-b3c4-93ca07625cb0 req-31e78353-894e-42ac-ba6a-17f3bd7edf74 service nova] Releasing lock "refresh_cache-79a4d3c8-ead6-4893-b0a7-89a8ff435e54" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 667.060216] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Acquired lock "refresh_cache-79a4d3c8-ead6-4893-b0a7-89a8ff435e54" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.060216] env[62965]: DEBUG nova.network.neutron [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 667.396820] env[62965]: ERROR nova.compute.manager [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 85c8faec-56d2-4fad-86ff-cc4710a93278, please check neutron logs for more information. [ 667.396820] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 667.396820] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 667.396820] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 667.396820] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 667.396820] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 667.396820] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 667.396820] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 667.396820] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 667.396820] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 667.396820] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 667.396820] env[62965]: ERROR nova.compute.manager raise self.value [ 667.396820] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 667.396820] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 667.396820] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 667.396820] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 667.397565] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 667.397565] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 667.397565] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 85c8faec-56d2-4fad-86ff-cc4710a93278, please check neutron logs for more information. [ 667.397565] env[62965]: ERROR nova.compute.manager [ 667.397565] env[62965]: Traceback (most recent call last): [ 667.397565] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 667.397565] env[62965]: listener.cb(fileno) [ 667.397565] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 667.397565] env[62965]: result = function(*args, **kwargs) [ 667.397565] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 667.397565] env[62965]: return func(*args, **kwargs) [ 667.397565] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 667.397565] env[62965]: raise e [ 667.397565] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 667.397565] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 667.397565] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 667.397565] env[62965]: created_port_ids = self._update_ports_for_instance( [ 667.397565] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 667.397565] env[62965]: with excutils.save_and_reraise_exception(): [ 667.397565] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 667.397565] env[62965]: self.force_reraise() [ 667.397565] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 667.397565] env[62965]: raise self.value [ 667.397565] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 667.397565] env[62965]: updated_port = self._update_port( [ 667.397565] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 667.397565] env[62965]: _ensure_no_port_binding_failure(port) [ 667.397565] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 667.397565] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 667.398551] env[62965]: nova.exception.PortBindingFailed: Binding failed for port 85c8faec-56d2-4fad-86ff-cc4710a93278, please check neutron logs for more information. [ 667.398551] env[62965]: Removing descriptor: 19 [ 667.406339] env[62965]: DEBUG nova.compute.manager [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 667.444755] env[62965]: DEBUG nova.virt.hardware [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 667.444973] env[62965]: DEBUG nova.virt.hardware [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 667.445148] env[62965]: DEBUG nova.virt.hardware [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 667.445434] env[62965]: DEBUG nova.virt.hardware [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 667.445514] env[62965]: DEBUG nova.virt.hardware [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 667.445612] env[62965]: DEBUG nova.virt.hardware [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 667.445818] env[62965]: DEBUG nova.virt.hardware [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 667.445975] env[62965]: DEBUG nova.virt.hardware [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 667.450281] env[62965]: DEBUG nova.virt.hardware [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 667.450585] env[62965]: DEBUG nova.virt.hardware [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 667.450651] env[62965]: DEBUG nova.virt.hardware [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 667.451531] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9128f182-fa04-4052-aa43-62837bc0cfb0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.460446] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caa0fd9d-da2d-4887-a857-bec4eafb10c5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.476226] env[62965]: ERROR nova.compute.manager [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 85c8faec-56d2-4fad-86ff-cc4710a93278, please check neutron logs for more information. [ 667.476226] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Traceback (most recent call last): [ 667.476226] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 667.476226] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] yield resources [ 667.476226] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 667.476226] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] self.driver.spawn(context, instance, image_meta, [ 667.476226] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 667.476226] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 667.476226] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 667.476226] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] vm_ref = self.build_virtual_machine(instance, [ 667.476226] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 667.476590] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] vif_infos = vmwarevif.get_vif_info(self._session, [ 667.476590] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 667.476590] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] for vif in network_info: [ 667.476590] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 667.476590] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] return self._sync_wrapper(fn, *args, **kwargs) [ 667.476590] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 667.476590] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] self.wait() [ 667.476590] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 667.476590] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] self[:] = self._gt.wait() [ 667.476590] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 667.476590] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] return self._exit_event.wait() [ 667.476590] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 667.476590] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] current.throw(*self._exc) [ 667.476916] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 667.476916] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] result = function(*args, **kwargs) [ 667.476916] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 667.476916] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] return func(*args, **kwargs) [ 667.476916] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 667.476916] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] raise e [ 667.476916] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 667.476916] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] nwinfo = self.network_api.allocate_for_instance( [ 667.476916] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 667.476916] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] created_port_ids = self._update_ports_for_instance( [ 667.476916] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 667.476916] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] with excutils.save_and_reraise_exception(): [ 667.476916] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 667.477250] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] self.force_reraise() [ 667.477250] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 667.477250] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] raise self.value [ 667.477250] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 667.477250] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] updated_port = self._update_port( [ 667.477250] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 667.477250] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] _ensure_no_port_binding_failure(port) [ 667.477250] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 667.477250] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] raise exception.PortBindingFailed(port_id=port['id']) [ 667.477250] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] nova.exception.PortBindingFailed: Binding failed for port 85c8faec-56d2-4fad-86ff-cc4710a93278, please check neutron logs for more information. [ 667.477250] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] [ 667.477250] env[62965]: INFO nova.compute.manager [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Terminating instance [ 667.485057] env[62965]: DEBUG oslo_concurrency.lockutils [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.102s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.486377] env[62965]: ERROR nova.compute.manager [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d0654b21-e661-454c-8b61-f9b30be9a633, please check neutron logs for more information. [ 667.486377] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Traceback (most recent call last): [ 667.486377] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 667.486377] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] self.driver.spawn(context, instance, image_meta, [ 667.486377] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 667.486377] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] self._vmops.spawn(context, instance, image_meta, injected_files, [ 667.486377] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 667.486377] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] vm_ref = self.build_virtual_machine(instance, [ 667.486377] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 667.486377] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] vif_infos = vmwarevif.get_vif_info(self._session, [ 667.486377] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 667.486858] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] for vif in network_info: [ 667.486858] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 667.486858] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] return self._sync_wrapper(fn, *args, **kwargs) [ 667.486858] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 667.486858] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] self.wait() [ 667.486858] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 667.486858] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] self[:] = self._gt.wait() [ 667.486858] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 667.486858] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] return self._exit_event.wait() [ 667.486858] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 667.486858] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] result = hub.switch() [ 667.486858] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 667.486858] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] return self.greenlet.switch() [ 667.487332] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 667.487332] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] result = function(*args, **kwargs) [ 667.487332] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 667.487332] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] return func(*args, **kwargs) [ 667.487332] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 667.487332] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] raise e [ 667.487332] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 667.487332] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] nwinfo = self.network_api.allocate_for_instance( [ 667.487332] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 667.487332] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] created_port_ids = self._update_ports_for_instance( [ 667.487332] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 667.487332] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] with excutils.save_and_reraise_exception(): [ 667.487332] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 667.487811] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] self.force_reraise() [ 667.487811] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 667.487811] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] raise self.value [ 667.487811] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 667.487811] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] updated_port = self._update_port( [ 667.487811] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 667.487811] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] _ensure_no_port_binding_failure(port) [ 667.487811] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 667.487811] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] raise exception.PortBindingFailed(port_id=port['id']) [ 667.487811] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] nova.exception.PortBindingFailed: Binding failed for port d0654b21-e661-454c-8b61-f9b30be9a633, please check neutron logs for more information. [ 667.487811] env[62965]: ERROR nova.compute.manager [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] [ 667.488573] env[62965]: DEBUG nova.compute.utils [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Binding failed for port d0654b21-e661-454c-8b61-f9b30be9a633, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 667.488573] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.239s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 667.490745] env[62965]: INFO nova.compute.claims [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 667.495584] env[62965]: DEBUG nova.compute.manager [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Build of instance 222f547e-8fce-4576-bafc-809ea2ee8631 was re-scheduled: Binding failed for port d0654b21-e661-454c-8b61-f9b30be9a633, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 667.496034] env[62965]: DEBUG nova.compute.manager [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 667.496266] env[62965]: DEBUG oslo_concurrency.lockutils [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] Acquiring lock "refresh_cache-222f547e-8fce-4576-bafc-809ea2ee8631" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 667.496414] env[62965]: DEBUG oslo_concurrency.lockutils [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] Acquired lock "refresh_cache-222f547e-8fce-4576-bafc-809ea2ee8631" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.496571] env[62965]: DEBUG nova.network.neutron [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 667.581623] env[62965]: DEBUG nova.network.neutron [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 667.659171] env[62965]: DEBUG nova.network.neutron [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.687414] env[62965]: DEBUG nova.compute.manager [req-35878274-ac33-47c1-b2fd-b10b8368976e req-7c4ad9d3-b2bf-4d61-bab9-9e251aab12bd service nova] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Received event network-changed-85c8faec-56d2-4fad-86ff-cc4710a93278 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 667.687414] env[62965]: DEBUG nova.compute.manager [req-35878274-ac33-47c1-b2fd-b10b8368976e req-7c4ad9d3-b2bf-4d61-bab9-9e251aab12bd service nova] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Refreshing instance network info cache due to event network-changed-85c8faec-56d2-4fad-86ff-cc4710a93278. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 667.687414] env[62965]: DEBUG oslo_concurrency.lockutils [req-35878274-ac33-47c1-b2fd-b10b8368976e req-7c4ad9d3-b2bf-4d61-bab9-9e251aab12bd service nova] Acquiring lock "refresh_cache-7eee167f-50e3-48d5-9c6d-9115df9bb0b9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 667.687414] env[62965]: DEBUG oslo_concurrency.lockutils [req-35878274-ac33-47c1-b2fd-b10b8368976e req-7c4ad9d3-b2bf-4d61-bab9-9e251aab12bd service nova] Acquired lock "refresh_cache-7eee167f-50e3-48d5-9c6d-9115df9bb0b9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.687414] env[62965]: DEBUG nova.network.neutron [req-35878274-ac33-47c1-b2fd-b10b8368976e req-7c4ad9d3-b2bf-4d61-bab9-9e251aab12bd service nova] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Refreshing network info cache for port 85c8faec-56d2-4fad-86ff-cc4710a93278 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 667.933489] env[62965]: DEBUG nova.compute.manager [req-09231f40-960a-4081-aad1-e118e682c43b req-f26adcfc-e7b6-486f-9ef1-78454c6d3c77 service nova] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Received event network-vif-deleted-76d8df8a-8e64-4dd2-b6bd-f1726c488d7f {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 667.981555] env[62965]: DEBUG oslo_concurrency.lockutils [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Acquiring lock "refresh_cache-7eee167f-50e3-48d5-9c6d-9115df9bb0b9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 668.017977] env[62965]: DEBUG nova.network.neutron [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 668.094318] env[62965]: DEBUG nova.network.neutron [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.165128] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Releasing lock "refresh_cache-79a4d3c8-ead6-4893-b0a7-89a8ff435e54" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 668.165570] env[62965]: DEBUG nova.compute.manager [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 668.165763] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 668.166100] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-df754f43-4736-4158-bff0-4ea8c8e1ff95 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.174771] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b870337-5cbe-4f47-8ae7-f2a7cfcb200b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.200316] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 79a4d3c8-ead6-4893-b0a7-89a8ff435e54 could not be found. [ 668.200566] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 668.200792] env[62965]: INFO nova.compute.manager [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Took 0.03 seconds to destroy the instance on the hypervisor. [ 668.201185] env[62965]: DEBUG oslo.service.loopingcall [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 668.201942] env[62965]: DEBUG nova.compute.manager [-] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 668.202122] env[62965]: DEBUG nova.network.neutron [-] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 668.208873] env[62965]: DEBUG nova.network.neutron [req-35878274-ac33-47c1-b2fd-b10b8368976e req-7c4ad9d3-b2bf-4d61-bab9-9e251aab12bd service nova] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 668.220263] env[62965]: DEBUG nova.network.neutron [-] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 668.324723] env[62965]: DEBUG nova.network.neutron [req-35878274-ac33-47c1-b2fd-b10b8368976e req-7c4ad9d3-b2bf-4d61-bab9-9e251aab12bd service nova] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.600574] env[62965]: DEBUG oslo_concurrency.lockutils [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] Releasing lock "refresh_cache-222f547e-8fce-4576-bafc-809ea2ee8631" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 668.600812] env[62965]: DEBUG nova.compute.manager [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 668.600991] env[62965]: DEBUG nova.compute.manager [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 668.601204] env[62965]: DEBUG nova.network.neutron [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 668.625828] env[62965]: DEBUG nova.network.neutron [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 668.725429] env[62965]: DEBUG nova.network.neutron [-] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.830189] env[62965]: DEBUG oslo_concurrency.lockutils [req-35878274-ac33-47c1-b2fd-b10b8368976e req-7c4ad9d3-b2bf-4d61-bab9-9e251aab12bd service nova] Releasing lock "refresh_cache-7eee167f-50e3-48d5-9c6d-9115df9bb0b9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 668.830541] env[62965]: DEBUG nova.compute.manager [req-35878274-ac33-47c1-b2fd-b10b8368976e req-7c4ad9d3-b2bf-4d61-bab9-9e251aab12bd service nova] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Received event network-vif-deleted-85c8faec-56d2-4fad-86ff-cc4710a93278 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 668.830923] env[62965]: DEBUG oslo_concurrency.lockutils [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Acquired lock "refresh_cache-7eee167f-50e3-48d5-9c6d-9115df9bb0b9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.831706] env[62965]: DEBUG nova.network.neutron [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 668.972034] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91085693-0952-4e21-ae46-86bea88ea2f2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.980778] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-696fb0d7-53fc-4852-8afd-194c364aa57b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.013800] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2344dab3-a429-492d-998e-964b47a123fd {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.021512] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55e96d8b-d73f-4781-a717-2482a7eb33db {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.036374] env[62965]: DEBUG nova.compute.provider_tree [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 669.128172] env[62965]: DEBUG nova.network.neutron [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.229831] env[62965]: INFO nova.compute.manager [-] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Took 1.03 seconds to deallocate network for instance. [ 669.232332] env[62965]: DEBUG nova.compute.claims [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 669.232332] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 669.353956] env[62965]: DEBUG nova.network.neutron [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 669.419212] env[62965]: DEBUG nova.network.neutron [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.540099] env[62965]: DEBUG nova.scheduler.client.report [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 669.631301] env[62965]: INFO nova.compute.manager [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] [instance: 222f547e-8fce-4576-bafc-809ea2ee8631] Took 1.03 seconds to deallocate network for instance. [ 669.922615] env[62965]: DEBUG oslo_concurrency.lockutils [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Releasing lock "refresh_cache-7eee167f-50e3-48d5-9c6d-9115df9bb0b9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 669.923064] env[62965]: DEBUG nova.compute.manager [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 669.923268] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 669.924459] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f09ae235-be8b-4377-a29f-5d5b3d484e1c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.932738] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57727891-0f51-4a23-b2c4-2b58a8f1f852 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.955231] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7eee167f-50e3-48d5-9c6d-9115df9bb0b9 could not be found. [ 669.955454] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 669.955631] env[62965]: INFO nova.compute.manager [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Took 0.03 seconds to destroy the instance on the hypervisor. [ 669.955870] env[62965]: DEBUG oslo.service.loopingcall [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 669.956116] env[62965]: DEBUG nova.compute.manager [-] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 669.956206] env[62965]: DEBUG nova.network.neutron [-] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 669.971382] env[62965]: DEBUG nova.network.neutron [-] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 670.046274] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.558s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.046807] env[62965]: DEBUG nova.compute.manager [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 670.050103] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.774s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.398561] env[62965]: DEBUG oslo_concurrency.lockutils [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] Acquiring lock "6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.398801] env[62965]: DEBUG oslo_concurrency.lockutils [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] Lock "6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.474487] env[62965]: DEBUG nova.network.neutron [-] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.554875] env[62965]: DEBUG nova.compute.utils [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 670.559258] env[62965]: DEBUG nova.compute.manager [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 670.559490] env[62965]: DEBUG nova.network.neutron [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 670.607992] env[62965]: DEBUG nova.policy [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c584837dbebe4491af4a2e12d7adb273', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5f656d82f77d45c5a5b55673a2d01cb3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 670.662256] env[62965]: INFO nova.scheduler.client.report [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] Deleted allocations for instance 222f547e-8fce-4576-bafc-809ea2ee8631 [ 670.979905] env[62965]: INFO nova.compute.manager [-] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Took 1.02 seconds to deallocate network for instance. [ 670.983282] env[62965]: DEBUG nova.compute.claims [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 670.983450] env[62965]: DEBUG oslo_concurrency.lockutils [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 671.001011] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92dd7c54-fea9-4e84-af65-c7d272a5b940 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.009066] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf247790-38c0-4f1d-ad7c-1183aa8d138b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.043775] env[62965]: DEBUG nova.network.neutron [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Successfully created port: dacdb675-46d2-428a-9408-131ed8d7743b {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 671.046406] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79f922a1-fa6b-4535-b43c-f6e14896203a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.054995] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea96ed02-ffc4-4cfe-85bc-3ee0101fa3e2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.059581] env[62965]: DEBUG nova.compute.manager [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 671.073572] env[62965]: DEBUG nova.compute.provider_tree [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 671.174372] env[62965]: DEBUG oslo_concurrency.lockutils [None req-756a80d4-ecc4-4baf-97d2-515e3fe3cb5e tempest-FloatingIPsAssociationTestJSON-1139767551 tempest-FloatingIPsAssociationTestJSON-1139767551-project-member] Lock "222f547e-8fce-4576-bafc-809ea2ee8631" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.822s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 671.575726] env[62965]: DEBUG nova.scheduler.client.report [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 671.677548] env[62965]: DEBUG nova.compute.manager [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 671.779369] env[62965]: DEBUG nova.compute.manager [req-8ea50d9f-8459-40f9-a7ab-6bdca52621fc req-edd9ef9a-8f8d-4513-ac1d-81824ce4a646 service nova] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Received event network-changed-dacdb675-46d2-428a-9408-131ed8d7743b {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 671.779369] env[62965]: DEBUG nova.compute.manager [req-8ea50d9f-8459-40f9-a7ab-6bdca52621fc req-edd9ef9a-8f8d-4513-ac1d-81824ce4a646 service nova] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Refreshing instance network info cache due to event network-changed-dacdb675-46d2-428a-9408-131ed8d7743b. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 671.779369] env[62965]: DEBUG oslo_concurrency.lockutils [req-8ea50d9f-8459-40f9-a7ab-6bdca52621fc req-edd9ef9a-8f8d-4513-ac1d-81824ce4a646 service nova] Acquiring lock "refresh_cache-f45f605f-cb91-4169-879c-bf9e6ce9c989" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.779369] env[62965]: DEBUG oslo_concurrency.lockutils [req-8ea50d9f-8459-40f9-a7ab-6bdca52621fc req-edd9ef9a-8f8d-4513-ac1d-81824ce4a646 service nova] Acquired lock "refresh_cache-f45f605f-cb91-4169-879c-bf9e6ce9c989" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.779369] env[62965]: DEBUG nova.network.neutron [req-8ea50d9f-8459-40f9-a7ab-6bdca52621fc req-edd9ef9a-8f8d-4513-ac1d-81824ce4a646 service nova] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Refreshing network info cache for port dacdb675-46d2-428a-9408-131ed8d7743b {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 671.948426] env[62965]: ERROR nova.compute.manager [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port dacdb675-46d2-428a-9408-131ed8d7743b, please check neutron logs for more information. [ 671.948426] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 671.948426] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 671.948426] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 671.948426] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 671.948426] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 671.948426] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 671.948426] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 671.948426] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 671.948426] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 671.948426] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 671.948426] env[62965]: ERROR nova.compute.manager raise self.value [ 671.948426] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 671.948426] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 671.948426] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 671.948426] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 671.948928] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 671.948928] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 671.948928] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port dacdb675-46d2-428a-9408-131ed8d7743b, please check neutron logs for more information. [ 671.948928] env[62965]: ERROR nova.compute.manager [ 671.948928] env[62965]: Traceback (most recent call last): [ 671.948928] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 671.948928] env[62965]: listener.cb(fileno) [ 671.948928] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 671.948928] env[62965]: result = function(*args, **kwargs) [ 671.948928] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 671.948928] env[62965]: return func(*args, **kwargs) [ 671.948928] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 671.948928] env[62965]: raise e [ 671.948928] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 671.948928] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 671.948928] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 671.948928] env[62965]: created_port_ids = self._update_ports_for_instance( [ 671.948928] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 671.948928] env[62965]: with excutils.save_and_reraise_exception(): [ 671.948928] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 671.948928] env[62965]: self.force_reraise() [ 671.948928] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 671.948928] env[62965]: raise self.value [ 671.948928] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 671.948928] env[62965]: updated_port = self._update_port( [ 671.948928] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 671.948928] env[62965]: _ensure_no_port_binding_failure(port) [ 671.948928] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 671.948928] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 671.949782] env[62965]: nova.exception.PortBindingFailed: Binding failed for port dacdb675-46d2-428a-9408-131ed8d7743b, please check neutron logs for more information. [ 671.949782] env[62965]: Removing descriptor: 19 [ 672.073706] env[62965]: DEBUG nova.compute.manager [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 672.082953] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.030s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 672.082953] env[62965]: ERROR nova.compute.manager [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f0cea45c-6193-4d2f-8a27-0fee3faf6256, please check neutron logs for more information. [ 672.082953] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Traceback (most recent call last): [ 672.082953] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 672.082953] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] self.driver.spawn(context, instance, image_meta, [ 672.082953] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 672.082953] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 672.082953] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 672.082953] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] vm_ref = self.build_virtual_machine(instance, [ 672.083398] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 672.083398] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] vif_infos = vmwarevif.get_vif_info(self._session, [ 672.083398] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 672.083398] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] for vif in network_info: [ 672.083398] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 672.083398] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] return self._sync_wrapper(fn, *args, **kwargs) [ 672.083398] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 672.083398] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] self.wait() [ 672.083398] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 672.083398] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] self[:] = self._gt.wait() [ 672.083398] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 672.083398] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] return self._exit_event.wait() [ 672.083398] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 672.083756] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] result = hub.switch() [ 672.083756] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 672.083756] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] return self.greenlet.switch() [ 672.083756] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 672.083756] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] result = function(*args, **kwargs) [ 672.083756] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 672.083756] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] return func(*args, **kwargs) [ 672.083756] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 672.083756] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] raise e [ 672.083756] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 672.083756] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] nwinfo = self.network_api.allocate_for_instance( [ 672.083756] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 672.083756] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] created_port_ids = self._update_ports_for_instance( [ 672.084273] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 672.084273] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] with excutils.save_and_reraise_exception(): [ 672.084273] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.084273] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] self.force_reraise() [ 672.084273] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.084273] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] raise self.value [ 672.084273] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 672.084273] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] updated_port = self._update_port( [ 672.084273] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 672.084273] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] _ensure_no_port_binding_failure(port) [ 672.084273] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 672.084273] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] raise exception.PortBindingFailed(port_id=port['id']) [ 672.084616] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] nova.exception.PortBindingFailed: Binding failed for port f0cea45c-6193-4d2f-8a27-0fee3faf6256, please check neutron logs for more information. [ 672.084616] env[62965]: ERROR nova.compute.manager [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] [ 672.084616] env[62965]: DEBUG nova.compute.utils [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Binding failed for port f0cea45c-6193-4d2f-8a27-0fee3faf6256, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 672.087167] env[62965]: DEBUG nova.compute.manager [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Build of instance af29af38-bd31-446a-a944-2bb43d9aa6c6 was re-scheduled: Binding failed for port f0cea45c-6193-4d2f-8a27-0fee3faf6256, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 672.087167] env[62965]: DEBUG nova.compute.manager [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 672.087167] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Acquiring lock "refresh_cache-af29af38-bd31-446a-a944-2bb43d9aa6c6" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 672.087167] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Acquired lock "refresh_cache-af29af38-bd31-446a-a944-2bb43d9aa6c6" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.087597] env[62965]: DEBUG nova.network.neutron [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 672.088934] env[62965]: DEBUG oslo_concurrency.lockutils [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.650s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 672.102716] env[62965]: DEBUG nova.virt.hardware [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 672.103086] env[62965]: DEBUG nova.virt.hardware [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 672.103382] env[62965]: DEBUG nova.virt.hardware [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 672.104839] env[62965]: DEBUG nova.virt.hardware [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 672.104839] env[62965]: DEBUG nova.virt.hardware [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 672.104839] env[62965]: DEBUG nova.virt.hardware [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 672.104839] env[62965]: DEBUG nova.virt.hardware [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 672.104839] env[62965]: DEBUG nova.virt.hardware [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 672.105074] env[62965]: DEBUG nova.virt.hardware [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 672.105074] env[62965]: DEBUG nova.virt.hardware [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 672.105074] env[62965]: DEBUG nova.virt.hardware [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 672.106031] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf746b6c-3552-417e-a961-ce2b118acf42 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.118952] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-723765ed-6932-4436-b7c6-7eadc23238fa {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.137013] env[62965]: ERROR nova.compute.manager [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port dacdb675-46d2-428a-9408-131ed8d7743b, please check neutron logs for more information. [ 672.137013] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Traceback (most recent call last): [ 672.137013] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 672.137013] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] yield resources [ 672.137013] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 672.137013] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] self.driver.spawn(context, instance, image_meta, [ 672.137013] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 672.137013] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] self._vmops.spawn(context, instance, image_meta, injected_files, [ 672.137013] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 672.137013] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] vm_ref = self.build_virtual_machine(instance, [ 672.137013] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 672.137511] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] vif_infos = vmwarevif.get_vif_info(self._session, [ 672.137511] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 672.137511] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] for vif in network_info: [ 672.137511] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 672.137511] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] return self._sync_wrapper(fn, *args, **kwargs) [ 672.137511] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 672.137511] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] self.wait() [ 672.137511] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 672.137511] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] self[:] = self._gt.wait() [ 672.137511] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 672.137511] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] return self._exit_event.wait() [ 672.137511] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 672.137511] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] current.throw(*self._exc) [ 672.137892] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 672.137892] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] result = function(*args, **kwargs) [ 672.137892] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 672.137892] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] return func(*args, **kwargs) [ 672.137892] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 672.137892] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] raise e [ 672.137892] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 672.137892] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] nwinfo = self.network_api.allocate_for_instance( [ 672.137892] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 672.137892] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] created_port_ids = self._update_ports_for_instance( [ 672.137892] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 672.137892] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] with excutils.save_and_reraise_exception(): [ 672.137892] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.138278] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] self.force_reraise() [ 672.138278] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.138278] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] raise self.value [ 672.138278] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 672.138278] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] updated_port = self._update_port( [ 672.138278] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 672.138278] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] _ensure_no_port_binding_failure(port) [ 672.138278] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 672.138278] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] raise exception.PortBindingFailed(port_id=port['id']) [ 672.138278] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] nova.exception.PortBindingFailed: Binding failed for port dacdb675-46d2-428a-9408-131ed8d7743b, please check neutron logs for more information. [ 672.138278] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] [ 672.138278] env[62965]: INFO nova.compute.manager [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Terminating instance [ 672.205750] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 672.302766] env[62965]: DEBUG nova.network.neutron [req-8ea50d9f-8459-40f9-a7ab-6bdca52621fc req-edd9ef9a-8f8d-4513-ac1d-81824ce4a646 service nova] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 672.384621] env[62965]: DEBUG nova.network.neutron [req-8ea50d9f-8459-40f9-a7ab-6bdca52621fc req-edd9ef9a-8f8d-4513-ac1d-81824ce4a646 service nova] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.617794] env[62965]: DEBUG nova.network.neutron [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 672.641021] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] Acquiring lock "refresh_cache-f45f605f-cb91-4169-879c-bf9e6ce9c989" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 672.716351] env[62965]: DEBUG nova.network.neutron [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.886864] env[62965]: DEBUG oslo_concurrency.lockutils [req-8ea50d9f-8459-40f9-a7ab-6bdca52621fc req-edd9ef9a-8f8d-4513-ac1d-81824ce4a646 service nova] Releasing lock "refresh_cache-f45f605f-cb91-4169-879c-bf9e6ce9c989" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.887306] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] Acquired lock "refresh_cache-f45f605f-cb91-4169-879c-bf9e6ce9c989" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.887485] env[62965]: DEBUG nova.network.neutron [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 673.057168] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef8f4ab2-4156-46ef-bdd1-9da384535d56 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.064670] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9031799d-b995-4d26-86b8-e6da691e1002 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.095193] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27578893-6026-48db-9501-0ba8f7d9639d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.102278] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e637229e-025f-45bc-894f-7330717eea84 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.115660] env[62965]: DEBUG nova.compute.provider_tree [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 673.220975] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Releasing lock "refresh_cache-af29af38-bd31-446a-a944-2bb43d9aa6c6" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 673.221241] env[62965]: DEBUG nova.compute.manager [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 673.221428] env[62965]: DEBUG nova.compute.manager [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 673.221598] env[62965]: DEBUG nova.network.neutron [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 673.237447] env[62965]: DEBUG nova.network.neutron [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 673.407654] env[62965]: DEBUG nova.network.neutron [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 673.512443] env[62965]: DEBUG nova.network.neutron [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.620107] env[62965]: DEBUG nova.scheduler.client.report [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 673.740701] env[62965]: DEBUG nova.network.neutron [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.815907] env[62965]: DEBUG nova.compute.manager [req-71e4f978-f2e8-4dc1-bbe6-d486b62ec3fd req-ac56084e-87a6-4dd2-8dbe-69cff886d28a service nova] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Received event network-vif-deleted-dacdb675-46d2-428a-9408-131ed8d7743b {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 674.015136] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] Releasing lock "refresh_cache-f45f605f-cb91-4169-879c-bf9e6ce9c989" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 674.015605] env[62965]: DEBUG nova.compute.manager [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 674.015756] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 674.016059] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-81347da8-70a4-4e7a-adb5-76b6665ebf01 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.025445] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f15e4345-70fd-46f2-b10f-3685db1b1cb5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.050452] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f45f605f-cb91-4169-879c-bf9e6ce9c989 could not be found. [ 674.050661] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 674.050846] env[62965]: INFO nova.compute.manager [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Took 0.04 seconds to destroy the instance on the hypervisor. [ 674.051106] env[62965]: DEBUG oslo.service.loopingcall [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 674.051332] env[62965]: DEBUG nova.compute.manager [-] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 674.051427] env[62965]: DEBUG nova.network.neutron [-] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 674.073645] env[62965]: DEBUG nova.network.neutron [-] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 674.126594] env[62965]: DEBUG oslo_concurrency.lockutils [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.038s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.127257] env[62965]: ERROR nova.compute.manager [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6764b048-89fb-436a-a183-ce56c1079e65, please check neutron logs for more information. [ 674.127257] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Traceback (most recent call last): [ 674.127257] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 674.127257] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] self.driver.spawn(context, instance, image_meta, [ 674.127257] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 674.127257] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] self._vmops.spawn(context, instance, image_meta, injected_files, [ 674.127257] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 674.127257] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] vm_ref = self.build_virtual_machine(instance, [ 674.127257] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 674.127257] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] vif_infos = vmwarevif.get_vif_info(self._session, [ 674.127257] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 674.127561] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] for vif in network_info: [ 674.127561] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 674.127561] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] return self._sync_wrapper(fn, *args, **kwargs) [ 674.127561] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 674.127561] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] self.wait() [ 674.127561] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 674.127561] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] self[:] = self._gt.wait() [ 674.127561] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 674.127561] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] return self._exit_event.wait() [ 674.127561] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 674.127561] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] result = hub.switch() [ 674.127561] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 674.127561] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] return self.greenlet.switch() [ 674.128053] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 674.128053] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] result = function(*args, **kwargs) [ 674.128053] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 674.128053] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] return func(*args, **kwargs) [ 674.128053] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 674.128053] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] raise e [ 674.128053] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 674.128053] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] nwinfo = self.network_api.allocate_for_instance( [ 674.128053] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 674.128053] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] created_port_ids = self._update_ports_for_instance( [ 674.128053] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 674.128053] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] with excutils.save_and_reraise_exception(): [ 674.128053] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 674.128443] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] self.force_reraise() [ 674.128443] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 674.128443] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] raise self.value [ 674.128443] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 674.128443] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] updated_port = self._update_port( [ 674.128443] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 674.128443] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] _ensure_no_port_binding_failure(port) [ 674.128443] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 674.128443] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] raise exception.PortBindingFailed(port_id=port['id']) [ 674.128443] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] nova.exception.PortBindingFailed: Binding failed for port 6764b048-89fb-436a-a183-ce56c1079e65, please check neutron logs for more information. [ 674.128443] env[62965]: ERROR nova.compute.manager [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] [ 674.128772] env[62965]: DEBUG nova.compute.utils [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Binding failed for port 6764b048-89fb-436a-a183-ce56c1079e65, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 674.129180] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.948s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.130931] env[62965]: INFO nova.compute.claims [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 674.133475] env[62965]: DEBUG nova.compute.manager [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Build of instance 8a893e6e-37d4-4fd4-b0db-d6633117bdce was re-scheduled: Binding failed for port 6764b048-89fb-436a-a183-ce56c1079e65, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 674.133911] env[62965]: DEBUG nova.compute.manager [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 674.135206] env[62965]: DEBUG oslo_concurrency.lockutils [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Acquiring lock "refresh_cache-8a893e6e-37d4-4fd4-b0db-d6633117bdce" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 674.135206] env[62965]: DEBUG oslo_concurrency.lockutils [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Acquired lock "refresh_cache-8a893e6e-37d4-4fd4-b0db-d6633117bdce" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.135206] env[62965]: DEBUG nova.network.neutron [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 674.243276] env[62965]: INFO nova.compute.manager [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: af29af38-bd31-446a-a944-2bb43d9aa6c6] Took 1.02 seconds to deallocate network for instance. [ 674.575193] env[62965]: DEBUG nova.network.neutron [-] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.658110] env[62965]: DEBUG nova.network.neutron [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 674.738787] env[62965]: DEBUG nova.network.neutron [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.078793] env[62965]: INFO nova.compute.manager [-] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Took 1.03 seconds to deallocate network for instance. [ 675.081932] env[62965]: DEBUG nova.compute.claims [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 675.082286] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 675.242780] env[62965]: DEBUG oslo_concurrency.lockutils [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Releasing lock "refresh_cache-8a893e6e-37d4-4fd4-b0db-d6633117bdce" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 675.242780] env[62965]: DEBUG nova.compute.manager [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 675.242780] env[62965]: DEBUG nova.compute.manager [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 675.242780] env[62965]: DEBUG nova.network.neutron [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 675.258292] env[62965]: DEBUG nova.network.neutron [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 675.272575] env[62965]: INFO nova.scheduler.client.report [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Deleted allocations for instance af29af38-bd31-446a-a944-2bb43d9aa6c6 [ 675.495339] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d63d6f37-63fc-4d1b-8d58-e869ec1c71f8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.502764] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1a96026-83a2-48f8-8e82-c1696441adba {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.533722] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c3e5159-df46-4f45-a4b7-52725a4d512b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.540649] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96ce2bf6-b0da-446d-aa7b-0761a1a1566c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.554384] env[62965]: DEBUG nova.compute.provider_tree [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 675.761381] env[62965]: DEBUG nova.network.neutron [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.782984] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c4aacf45-0a79-4d78-b16b-1b3abd70a928 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Lock "af29af38-bd31-446a-a944-2bb43d9aa6c6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.342s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.057762] env[62965]: DEBUG nova.scheduler.client.report [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 676.263815] env[62965]: INFO nova.compute.manager [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] [instance: 8a893e6e-37d4-4fd4-b0db-d6633117bdce] Took 1.02 seconds to deallocate network for instance. [ 676.285365] env[62965]: DEBUG nova.compute.manager [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 676.562935] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.434s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.563483] env[62965]: DEBUG nova.compute.manager [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 676.566048] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 16.421s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 676.805181] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 677.070071] env[62965]: DEBUG nova.compute.utils [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 677.071545] env[62965]: DEBUG nova.compute.manager [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Not allocating networking since 'none' was specified. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 677.294618] env[62965]: INFO nova.scheduler.client.report [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Deleted allocations for instance 8a893e6e-37d4-4fd4-b0db-d6633117bdce [ 677.576632] env[62965]: DEBUG nova.compute.manager [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 677.597372] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 8ac4264d-7d75-4c38-b8da-e3d90ffe9363 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 677.597644] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 971f72f8-2686-4ced-a2a1-b7ab8a1ea024 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 677.597755] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance eaf9ca94-4f45-403e-bb06-e9a9adf58f38 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 677.597852] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 79a4d3c8-ead6-4893-b0a7-89a8ff435e54 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 677.597976] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 7eee167f-50e3-48d5-9c6d-9115df9bb0b9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 677.598109] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance f45f605f-cb91-4169-879c-bf9e6ce9c989 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 677.598240] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 02f113d2-f991-4268-b9a4-59e2e6bcf7a6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 677.804030] env[62965]: DEBUG oslo_concurrency.lockutils [None req-77e0049d-bf66-40d8-adf5-478262d58116 tempest-ServerRescueNegativeTestJSON-203051354 tempest-ServerRescueNegativeTestJSON-203051354-project-member] Lock "8a893e6e-37d4-4fd4-b0db-d6633117bdce" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 111.704s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.101485] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 4519d37a-f8e6-4a25-99a1-399925199a58 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 678.306299] env[62965]: DEBUG nova.compute.manager [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 678.589223] env[62965]: DEBUG nova.compute.manager [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 678.605248] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 678.616455] env[62965]: DEBUG nova.virt.hardware [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 678.616589] env[62965]: DEBUG nova.virt.hardware [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 678.616726] env[62965]: DEBUG nova.virt.hardware [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 678.616909] env[62965]: DEBUG nova.virt.hardware [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 678.617068] env[62965]: DEBUG nova.virt.hardware [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 678.617214] env[62965]: DEBUG nova.virt.hardware [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 678.617419] env[62965]: DEBUG nova.virt.hardware [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 678.617576] env[62965]: DEBUG nova.virt.hardware [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 678.617742] env[62965]: DEBUG nova.virt.hardware [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 678.617901] env[62965]: DEBUG nova.virt.hardware [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 678.618104] env[62965]: DEBUG nova.virt.hardware [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 678.618998] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ab1a417-5354-4b77-9fd3-d984a85aadb9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.628227] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62473776-4627-4e0e-8228-fb0757a3bdf2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.642819] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Instance VIF info [] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 678.648433] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Creating folder: Project (35470a210128494b8ffaa05d7ebab9ff). Parent ref: group-v294931. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 678.648679] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-318e5c76-b167-4898-926f-81e8125dd063 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.660142] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Created folder: Project (35470a210128494b8ffaa05d7ebab9ff) in parent group-v294931. [ 678.660330] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Creating folder: Instances. Parent ref: group-v294943. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 678.660556] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0a41ab48-57de-4dac-b30c-15e0621f040b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.669155] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Created folder: Instances in parent group-v294943. [ 678.669403] env[62965]: DEBUG oslo.service.loopingcall [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 678.669591] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 678.669784] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-91b9b203-ee50-4f87-926e-255938c79ffc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.686399] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 678.686399] env[62965]: value = "task-1389751" [ 678.686399] env[62965]: _type = "Task" [ 678.686399] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.694228] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389751, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.830525] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 679.108701] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 852e7cdf-5da3-4217-8960-f3204d4c1035 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 679.197426] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389751, 'name': CreateVM_Task, 'duration_secs': 0.271958} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.197616] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 679.198604] env[62965]: DEBUG oslo_vmware.service [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a9b4e7b-22f3-4c5c-860b-58574335b9f2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.205335] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.205505] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.206189] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 679.206412] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c488e56-7c77-476e-9fe0-9b0d33b06eb1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.212810] env[62965]: DEBUG oslo_vmware.api [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Waiting for the task: (returnval){ [ 679.212810] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52f77ab3-17fc-e091-e04c-8741f089b87d" [ 679.212810] env[62965]: _type = "Task" [ 679.212810] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.223890] env[62965]: DEBUG oslo_vmware.api [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52f77ab3-17fc-e091-e04c-8741f089b87d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.611774] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance b2f38e92-37a1-4e0b-9292-b20ea43a2038 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 679.724341] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 679.724615] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 679.724851] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.724995] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.725191] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 679.725450] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8215593e-0c10-4a4a-9638-f45aa2853b53 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.742915] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 679.743119] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 679.743910] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdcbb2e1-6988-43ba-a570-169457b61b89 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.750325] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88941b8a-5dfa-4a68-b80c-ab1e6e2e9e69 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.755795] env[62965]: DEBUG oslo_vmware.api [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Waiting for the task: (returnval){ [ 679.755795] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52ebe169-aa38-1131-f397-ace15989a6b6" [ 679.755795] env[62965]: _type = "Task" [ 679.755795] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.766102] env[62965]: DEBUG oslo_vmware.api [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52ebe169-aa38-1131-f397-ace15989a6b6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.116288] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 83c4b731-780a-4f3c-bbdd-b624a52365ef has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 680.269072] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Preparing fetch location {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 680.269386] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Creating directory with path [datastore2] vmware_temp/66786a71-8be4-4f38-8553-1e4537c546f5/523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 680.269642] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e1dd9486-a0e3-444d-8f7c-b0e2e76aa805 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.293705] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Created directory with path [datastore2] vmware_temp/66786a71-8be4-4f38-8553-1e4537c546f5/523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 680.293705] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Fetch image to [datastore2] vmware_temp/66786a71-8be4-4f38-8553-1e4537c546f5/523fe138-8131-44f9-96b3-87f923f60cb8/tmp-sparse.vmdk {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 680.293705] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Downloading image file data 523fe138-8131-44f9-96b3-87f923f60cb8 to [datastore2] vmware_temp/66786a71-8be4-4f38-8553-1e4537c546f5/523fe138-8131-44f9-96b3-87f923f60cb8/tmp-sparse.vmdk on the data store datastore2 {{(pid=62965) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 680.294676] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2576ccf-e96b-4f43-a64d-37f14730e59b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.302881] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8365a4a-e5ed-40be-9d86-04c67cf15c5d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.313424] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cef9dbc9-bfca-4a15-96a0-63f2c90f9f66 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.347897] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee3bf84e-25a0-4fc1-a75e-050924cf6f16 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.355528] env[62965]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-bbaae6dc-e417-4c7b-bbc4-1c0916acdb90 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.388440] env[62965]: DEBUG nova.virt.vmwareapi.images [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Downloading image file data 523fe138-8131-44f9-96b3-87f923f60cb8 to the data store datastore2 {{(pid=62965) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 680.444771] env[62965]: DEBUG oslo_vmware.rw_handles [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/66786a71-8be4-4f38-8553-1e4537c546f5/523fe138-8131-44f9-96b3-87f923f60cb8/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62965) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 680.619909] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance f8ba32e4-6234-4dee-8bc5-b7a12105a3c4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 681.122777] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance ece118fc-a39e-443a-bfd1-fc332da2452f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 681.193847] env[62965]: DEBUG oslo_vmware.rw_handles [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Completed reading data from the image iterator. {{(pid=62965) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 681.194082] env[62965]: DEBUG oslo_vmware.rw_handles [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Closing write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/66786a71-8be4-4f38-8553-1e4537c546f5/523fe138-8131-44f9-96b3-87f923f60cb8/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62965) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 681.249076] env[62965]: DEBUG nova.virt.vmwareapi.images [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Downloaded image file data 523fe138-8131-44f9-96b3-87f923f60cb8 to vmware_temp/66786a71-8be4-4f38-8553-1e4537c546f5/523fe138-8131-44f9-96b3-87f923f60cb8/tmp-sparse.vmdk on the data store datastore2 {{(pid=62965) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 681.251682] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Caching image {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 681.251987] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Copying Virtual Disk [datastore2] vmware_temp/66786a71-8be4-4f38-8553-1e4537c546f5/523fe138-8131-44f9-96b3-87f923f60cb8/tmp-sparse.vmdk to [datastore2] vmware_temp/66786a71-8be4-4f38-8553-1e4537c546f5/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 681.252309] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0d7f46c5-ad57-4e05-946c-c980a088169b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.261695] env[62965]: DEBUG oslo_vmware.api [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Waiting for the task: (returnval){ [ 681.261695] env[62965]: value = "task-1389754" [ 681.261695] env[62965]: _type = "Task" [ 681.261695] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.270713] env[62965]: DEBUG oslo_vmware.api [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389754, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.628748] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 681.773378] env[62965]: DEBUG oslo_vmware.api [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389754, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.131735] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 99db6bef-6501-4ab3-a329-0877ca303f1b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 682.273896] env[62965]: DEBUG oslo_vmware.api [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389754, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.633119} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.274098] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Copied Virtual Disk [datastore2] vmware_temp/66786a71-8be4-4f38-8553-1e4537c546f5/523fe138-8131-44f9-96b3-87f923f60cb8/tmp-sparse.vmdk to [datastore2] vmware_temp/66786a71-8be4-4f38-8553-1e4537c546f5/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 682.274283] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Deleting the datastore file [datastore2] vmware_temp/66786a71-8be4-4f38-8553-1e4537c546f5/523fe138-8131-44f9-96b3-87f923f60cb8/tmp-sparse.vmdk {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 682.274526] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f6430601-9c4b-4b85-9ea0-69dbb46f2e65 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.282713] env[62965]: DEBUG oslo_vmware.api [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Waiting for the task: (returnval){ [ 682.282713] env[62965]: value = "task-1389755" [ 682.282713] env[62965]: _type = "Task" [ 682.282713] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.291272] env[62965]: DEBUG oslo_vmware.api [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389755, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.635072] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 682.793153] env[62965]: DEBUG oslo_vmware.api [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389755, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.028529} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.793406] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 682.794845] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Moving file from [datastore2] vmware_temp/66786a71-8be4-4f38-8553-1e4537c546f5/523fe138-8131-44f9-96b3-87f923f60cb8 to [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8. {{(pid=62965) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 682.794845] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-7b3fdf87-dc34-47db-b84d-c4e83540daa5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.802493] env[62965]: DEBUG oslo_vmware.api [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Waiting for the task: (returnval){ [ 682.802493] env[62965]: value = "task-1389756" [ 682.802493] env[62965]: _type = "Task" [ 682.802493] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.811425] env[62965]: DEBUG oslo_vmware.api [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389756, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.139272] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance cd978d56-ba80-44a9-9a02-3a1254066571 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 683.316894] env[62965]: DEBUG oslo_vmware.api [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389756, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.030398} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.316990] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] File moved {{(pid=62965) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 683.317143] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Cleaning up location [datastore2] vmware_temp/66786a71-8be4-4f38-8553-1e4537c546f5 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 683.317304] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Deleting the datastore file [datastore2] vmware_temp/66786a71-8be4-4f38-8553-1e4537c546f5 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 683.317872] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-104e49c1-e77e-432c-ab4e-c979154a33af {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.327600] env[62965]: DEBUG oslo_vmware.api [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Waiting for the task: (returnval){ [ 683.327600] env[62965]: value = "task-1389757" [ 683.327600] env[62965]: _type = "Task" [ 683.327600] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.337726] env[62965]: DEBUG oslo_vmware.api [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389757, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.642821] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance f9a2b957-a312-4c9b-bb0d-0b1dad6416cd has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 683.701499] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] Acquiring lock "ab4da02e-f3ec-473b-968f-14b102780ddb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 683.702129] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] Lock "ab4da02e-f3ec-473b-968f-14b102780ddb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 683.838143] env[62965]: DEBUG oslo_vmware.api [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389757, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.026333} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.838450] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 683.839196] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f475b65c-8f01-4c13-bb59-895edbb589da {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.845060] env[62965]: DEBUG oslo_vmware.api [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Waiting for the task: (returnval){ [ 683.845060] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52c17a18-66b1-c170-a863-7ee14790f7c8" [ 683.845060] env[62965]: _type = "Task" [ 683.845060] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.854290] env[62965]: DEBUG oslo_vmware.api [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52c17a18-66b1-c170-a863-7ee14790f7c8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.146679] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 684.212268] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] Acquiring lock "c8402430-4229-4b61-9118-ddeab4d8cd79" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.212545] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] Lock "c8402430-4229-4b61-9118-ddeab4d8cd79" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.356443] env[62965]: DEBUG oslo_vmware.api [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52c17a18-66b1-c170-a863-7ee14790f7c8, 'name': SearchDatastore_Task, 'duration_secs': 0.010193} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.357355] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 684.357562] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 02f113d2-f991-4268-b9a4-59e2e6bcf7a6/02f113d2-f991-4268-b9a4-59e2e6bcf7a6.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 684.358568] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-efcba0dc-8e42-4700-9e49-b1a88feedb4d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.367121] env[62965]: DEBUG oslo_vmware.api [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Waiting for the task: (returnval){ [ 684.367121] env[62965]: value = "task-1389758" [ 684.367121] env[62965]: _type = "Task" [ 684.367121] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.376062] env[62965]: DEBUG oslo_vmware.api [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389758, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.649891] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 864ce643-8620-4405-b345-9164c81fec2a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 684.881206] env[62965]: DEBUG oslo_vmware.api [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389758, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.503792} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.881934] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 02f113d2-f991-4268-b9a4-59e2e6bcf7a6/02f113d2-f991-4268-b9a4-59e2e6bcf7a6.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 684.881934] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 684.882091] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e0a1d4f0-c7a6-405d-8df5-2fb4a5f21116 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.890039] env[62965]: DEBUG oslo_vmware.api [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Waiting for the task: (returnval){ [ 684.890039] env[62965]: value = "task-1389759" [ 684.890039] env[62965]: _type = "Task" [ 684.890039] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.898047] env[62965]: DEBUG oslo_vmware.api [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389759, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.156778] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance d6e90043-ebd3-4daf-acfa-545ce8462942 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 685.400774] env[62965]: DEBUG oslo_vmware.api [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389759, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069517} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.400774] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 685.400774] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a11e1b0-7c8c-4785-8cc8-44ee498f7e9e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.421115] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Reconfiguring VM instance instance-0000001f to attach disk [datastore2] 02f113d2-f991-4268-b9a4-59e2e6bcf7a6/02f113d2-f991-4268-b9a4-59e2e6bcf7a6.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 685.422023] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a6d8d716-00e1-4b67-a47b-20e0be1f714f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.445036] env[62965]: DEBUG oslo_vmware.api [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Waiting for the task: (returnval){ [ 685.445036] env[62965]: value = "task-1389760" [ 685.445036] env[62965]: _type = "Task" [ 685.445036] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.452438] env[62965]: DEBUG oslo_vmware.api [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389760, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.659759] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance afaec369-6aef-4bdc-b886-eed795644867 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 685.955608] env[62965]: DEBUG oslo_vmware.api [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389760, 'name': ReconfigVM_Task, 'duration_secs': 0.30389} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.955987] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Reconfigured VM instance instance-0000001f to attach disk [datastore2] 02f113d2-f991-4268-b9a4-59e2e6bcf7a6/02f113d2-f991-4268-b9a4-59e2e6bcf7a6.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 685.956658] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b5ca7c1a-e5bd-4a36-b7c2-e19355c8b6d2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.964090] env[62965]: DEBUG oslo_vmware.api [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Waiting for the task: (returnval){ [ 685.964090] env[62965]: value = "task-1389761" [ 685.964090] env[62965]: _type = "Task" [ 685.964090] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.974223] env[62965]: DEBUG oslo_vmware.api [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389761, 'name': Rename_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.162897] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 6e0a4e4b-21a2-4185-999e-69b4c74c37d7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 686.474458] env[62965]: DEBUG oslo_vmware.api [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389761, 'name': Rename_Task, 'duration_secs': 0.14691} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.474865] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 686.475319] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3361d113-932a-48da-b267-2794a6c6e984 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.482511] env[62965]: DEBUG oslo_vmware.api [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Waiting for the task: (returnval){ [ 686.482511] env[62965]: value = "task-1389762" [ 686.482511] env[62965]: _type = "Task" [ 686.482511] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.493381] env[62965]: DEBUG oslo_vmware.api [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389762, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.665474] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance e87adfae-3ea4-4cc4-9978-d8c811c80941 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 686.992436] env[62965]: DEBUG oslo_vmware.api [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389762, 'name': PowerOnVM_Task, 'duration_secs': 0.473957} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.992730] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 686.992922] env[62965]: INFO nova.compute.manager [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Took 8.40 seconds to spawn the instance on the hypervisor. [ 686.993118] env[62965]: DEBUG nova.compute.manager [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 686.993911] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b056dfb7-a80a-48c2-9ee7-4423f623c2e3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.168753] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 9a2a324c-4eb1-45b6-8414-42671b412429 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 687.510231] env[62965]: INFO nova.compute.manager [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Took 30.35 seconds to build instance. [ 687.675467] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance dcae9cae-d4f5-4b34-b392-00b5ef9c81bf has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 688.013181] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3db7c128-7b69-4d3f-88d2-c81063c505c2 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Lock "02f113d2-f991-4268-b9a4-59e2e6bcf7a6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 113.603s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.178675] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 688.179064] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Total usable vcpus: 48, total allocated vcpus: 7 {{(pid=62965) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 688.179260] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1920MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=7 pci_stats=[] {{(pid=62965) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 688.361710] env[62965]: INFO nova.compute.manager [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Rebuilding instance [ 688.400354] env[62965]: DEBUG nova.compute.manager [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 688.401381] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a27e8f6c-78cb-41a3-9216-79f4c326e113 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.511864] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23d01ff5-dd58-405d-a1e5-c969d98a2591 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.515969] env[62965]: DEBUG nova.compute.manager [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 688.522373] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c9f48e1-6a7f-4d39-9014-d1ef998e8908 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.552963] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1df294c6-1887-47f9-a357-be29a97b9a35 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.561911] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb62f5af-138a-453e-9798-489ea8ad0599 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.577031] env[62965]: DEBUG nova.compute.provider_tree [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 689.036711] env[62965]: DEBUG oslo_concurrency.lockutils [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 689.079384] env[62965]: DEBUG nova.scheduler.client.report [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 689.417438] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 689.417758] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b836d0fc-78c3-473b-8324-a911a63588d6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.426100] env[62965]: DEBUG oslo_vmware.api [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Waiting for the task: (returnval){ [ 689.426100] env[62965]: value = "task-1389763" [ 689.426100] env[62965]: _type = "Task" [ 689.426100] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.435156] env[62965]: DEBUG oslo_vmware.api [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389763, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.584364] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62965) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 689.584608] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 13.019s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 689.584881] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 28.558s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 689.938623] env[62965]: DEBUG oslo_vmware.api [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389763, 'name': PowerOffVM_Task, 'duration_secs': 0.119053} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.938917] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 689.939159] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 689.939969] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e64c173-ab0e-4f03-921d-936c471cc5df {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.947453] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 689.947709] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3a0b4661-911b-49af-89b3-8c8cf9432dd2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.978807] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 689.979022] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Deleting contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 689.979260] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Deleting the datastore file [datastore2] 02f113d2-f991-4268-b9a4-59e2e6bcf7a6 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 689.979529] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-71987c8b-a701-447e-a481-3c5d84d9ba23 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.987155] env[62965]: DEBUG oslo_vmware.api [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Waiting for the task: (returnval){ [ 689.987155] env[62965]: value = "task-1389765" [ 689.987155] env[62965]: _type = "Task" [ 689.987155] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.995308] env[62965]: DEBUG oslo_vmware.api [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389765, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.429571] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-168b57c0-0fe0-441a-8725-a3d46ab082b3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.437410] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-876b1da8-0a17-4aaf-b67b-c4faafcfad25 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.466664] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41671e9e-66de-4712-aa10-943129a14073 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.473987] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c868bfe8-d7ef-481a-989a-ef67ef0ede91 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.486976] env[62965]: DEBUG nova.compute.provider_tree [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 690.495319] env[62965]: DEBUG oslo_vmware.api [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389765, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.104814} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.495549] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 690.495732] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Deleted contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 690.495898] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 690.992127] env[62965]: DEBUG nova.scheduler.client.report [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 691.497195] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.912s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.497832] env[62965]: ERROR nova.compute.manager [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 763d2a98-a632-43dd-8f7e-f3c96720096e, please check neutron logs for more information. [ 691.497832] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Traceback (most recent call last): [ 691.497832] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 691.497832] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] self.driver.spawn(context, instance, image_meta, [ 691.497832] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 691.497832] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] self._vmops.spawn(context, instance, image_meta, injected_files, [ 691.497832] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 691.497832] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] vm_ref = self.build_virtual_machine(instance, [ 691.497832] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 691.497832] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] vif_infos = vmwarevif.get_vif_info(self._session, [ 691.497832] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 691.498226] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] for vif in network_info: [ 691.498226] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 691.498226] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] return self._sync_wrapper(fn, *args, **kwargs) [ 691.498226] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 691.498226] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] self.wait() [ 691.498226] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 691.498226] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] self[:] = self._gt.wait() [ 691.498226] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 691.498226] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] return self._exit_event.wait() [ 691.498226] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 691.498226] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] result = hub.switch() [ 691.498226] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 691.498226] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] return self.greenlet.switch() [ 691.498673] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 691.498673] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] result = function(*args, **kwargs) [ 691.498673] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 691.498673] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] return func(*args, **kwargs) [ 691.498673] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 691.498673] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] raise e [ 691.498673] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 691.498673] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] nwinfo = self.network_api.allocate_for_instance( [ 691.498673] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 691.498673] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] created_port_ids = self._update_ports_for_instance( [ 691.498673] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 691.498673] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] with excutils.save_and_reraise_exception(): [ 691.498673] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 691.499060] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] self.force_reraise() [ 691.499060] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 691.499060] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] raise self.value [ 691.499060] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 691.499060] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] updated_port = self._update_port( [ 691.499060] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 691.499060] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] _ensure_no_port_binding_failure(port) [ 691.499060] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 691.499060] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] raise exception.PortBindingFailed(port_id=port['id']) [ 691.499060] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] nova.exception.PortBindingFailed: Binding failed for port 763d2a98-a632-43dd-8f7e-f3c96720096e, please check neutron logs for more information. [ 691.499060] env[62965]: ERROR nova.compute.manager [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] [ 691.499447] env[62965]: DEBUG nova.compute.utils [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Binding failed for port 763d2a98-a632-43dd-8f7e-f3c96720096e, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 691.499809] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.660s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 691.504382] env[62965]: DEBUG nova.compute.manager [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Build of instance 8ac4264d-7d75-4c38-b8da-e3d90ffe9363 was re-scheduled: Binding failed for port 763d2a98-a632-43dd-8f7e-f3c96720096e, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 691.504701] env[62965]: DEBUG nova.compute.manager [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 691.504921] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] Acquiring lock "refresh_cache-8ac4264d-7d75-4c38-b8da-e3d90ffe9363" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 691.505083] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] Acquired lock "refresh_cache-8ac4264d-7d75-4c38-b8da-e3d90ffe9363" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.505242] env[62965]: DEBUG nova.network.neutron [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 691.531718] env[62965]: DEBUG nova.virt.hardware [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 691.531955] env[62965]: DEBUG nova.virt.hardware [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 691.532135] env[62965]: DEBUG nova.virt.hardware [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 691.532322] env[62965]: DEBUG nova.virt.hardware [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 691.532468] env[62965]: DEBUG nova.virt.hardware [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 691.532612] env[62965]: DEBUG nova.virt.hardware [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 691.532814] env[62965]: DEBUG nova.virt.hardware [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 691.532969] env[62965]: DEBUG nova.virt.hardware [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 691.533154] env[62965]: DEBUG nova.virt.hardware [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 691.533319] env[62965]: DEBUG nova.virt.hardware [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 691.533491] env[62965]: DEBUG nova.virt.hardware [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 691.534945] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-242f52b5-d3bd-4c84-af7c-b96dfb52c3c5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.543726] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d98ac6-5e3f-4c84-974c-1877cc6075e7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.557604] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Instance VIF info [] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 691.563116] env[62965]: DEBUG oslo.service.loopingcall [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 691.563352] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 691.563559] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3439fe67-4897-4303-8cf1-34763c3c0a5c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.580712] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 691.580712] env[62965]: value = "task-1389766" [ 691.580712] env[62965]: _type = "Task" [ 691.580712] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.588561] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389766, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.028584] env[62965]: DEBUG nova.network.neutron [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 692.092475] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389766, 'name': CreateVM_Task, 'duration_secs': 0.27452} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.092624] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 692.093034] env[62965]: DEBUG oslo_concurrency.lockutils [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 692.093197] env[62965]: DEBUG oslo_concurrency.lockutils [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.093529] env[62965]: DEBUG oslo_concurrency.lockutils [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 692.095712] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d658071a-6a3f-4e2e-8282-0ce96aa9cc19 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.100627] env[62965]: DEBUG oslo_vmware.api [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Waiting for the task: (returnval){ [ 692.100627] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]526404dc-051a-8747-2085-c93c5708cf58" [ 692.100627] env[62965]: _type = "Task" [ 692.100627] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.110539] env[62965]: DEBUG oslo_vmware.api [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]526404dc-051a-8747-2085-c93c5708cf58, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.112593] env[62965]: DEBUG nova.network.neutron [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.356703] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-046fe196-408f-4144-accb-ce9ad6e58b3f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.364628] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a287b36-cd3e-48b0-ac45-c854d158525b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.396982] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-704adcd3-df4e-40e4-a2e3-63c40db093d4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.405337] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecc98e46-b707-47f2-8ee5-5c3fd91c44ee {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.419865] env[62965]: DEBUG nova.compute.provider_tree [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 692.612332] env[62965]: DEBUG oslo_vmware.api [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]526404dc-051a-8747-2085-c93c5708cf58, 'name': SearchDatastore_Task, 'duration_secs': 0.458076} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.612332] env[62965]: DEBUG oslo_concurrency.lockutils [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 692.612332] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 692.612332] env[62965]: DEBUG oslo_concurrency.lockutils [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 692.612790] env[62965]: DEBUG oslo_concurrency.lockutils [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.612790] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 692.612790] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-48cc8ef2-c3c7-46b3-be4d-c8208094e53c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.614786] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] Releasing lock "refresh_cache-8ac4264d-7d75-4c38-b8da-e3d90ffe9363" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 692.614985] env[62965]: DEBUG nova.compute.manager [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 692.615183] env[62965]: DEBUG nova.compute.manager [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 692.615352] env[62965]: DEBUG nova.network.neutron [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 692.624292] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 692.624535] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 692.625395] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-652abe13-8aa5-4ed3-8f86-939f2e2ca228 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.631075] env[62965]: DEBUG nova.network.neutron [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 692.633387] env[62965]: DEBUG oslo_vmware.api [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Waiting for the task: (returnval){ [ 692.633387] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]528d16c8-0082-5ed1-9388-6ff345ec45c0" [ 692.633387] env[62965]: _type = "Task" [ 692.633387] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.642636] env[62965]: DEBUG oslo_vmware.api [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]528d16c8-0082-5ed1-9388-6ff345ec45c0, 'name': SearchDatastore_Task, 'duration_secs': 0.008469} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.643914] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b96cf56f-fcca-472c-84be-5465bbe66f54 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.649710] env[62965]: DEBUG oslo_vmware.api [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Waiting for the task: (returnval){ [ 692.649710] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5291ea0e-33ae-c116-9988-1c51ed6ee095" [ 692.649710] env[62965]: _type = "Task" [ 692.649710] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.657460] env[62965]: DEBUG oslo_vmware.api [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5291ea0e-33ae-c116-9988-1c51ed6ee095, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.922936] env[62965]: DEBUG nova.scheduler.client.report [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 693.138272] env[62965]: DEBUG nova.network.neutron [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 693.161025] env[62965]: DEBUG oslo_vmware.api [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5291ea0e-33ae-c116-9988-1c51ed6ee095, 'name': SearchDatastore_Task, 'duration_secs': 0.00818} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.161290] env[62965]: DEBUG oslo_concurrency.lockutils [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 693.161553] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] 02f113d2-f991-4268-b9a4-59e2e6bcf7a6/02f113d2-f991-4268-b9a4-59e2e6bcf7a6.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 693.161817] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-129cba90-c71d-4c24-81ba-e381c9f0f6e0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.168581] env[62965]: DEBUG oslo_vmware.api [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Waiting for the task: (returnval){ [ 693.168581] env[62965]: value = "task-1389767" [ 693.168581] env[62965]: _type = "Task" [ 693.168581] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.177167] env[62965]: DEBUG oslo_vmware.api [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389767, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.427907] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.928s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.428732] env[62965]: ERROR nova.compute.manager [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1da29949-9732-45b3-a088-7b379ed73ad5, please check neutron logs for more information. [ 693.428732] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Traceback (most recent call last): [ 693.428732] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 693.428732] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] self.driver.spawn(context, instance, image_meta, [ 693.428732] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 693.428732] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] self._vmops.spawn(context, instance, image_meta, injected_files, [ 693.428732] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 693.428732] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] vm_ref = self.build_virtual_machine(instance, [ 693.428732] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 693.428732] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] vif_infos = vmwarevif.get_vif_info(self._session, [ 693.428732] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 693.429137] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] for vif in network_info: [ 693.429137] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 693.429137] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] return self._sync_wrapper(fn, *args, **kwargs) [ 693.429137] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 693.429137] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] self.wait() [ 693.429137] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 693.429137] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] self[:] = self._gt.wait() [ 693.429137] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 693.429137] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] return self._exit_event.wait() [ 693.429137] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 693.429137] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] result = hub.switch() [ 693.429137] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 693.429137] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] return self.greenlet.switch() [ 693.429503] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 693.429503] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] result = function(*args, **kwargs) [ 693.429503] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 693.429503] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] return func(*args, **kwargs) [ 693.429503] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 693.429503] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] raise e [ 693.429503] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 693.429503] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] nwinfo = self.network_api.allocate_for_instance( [ 693.429503] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 693.429503] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] created_port_ids = self._update_ports_for_instance( [ 693.429503] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 693.429503] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] with excutils.save_and_reraise_exception(): [ 693.429503] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 693.429885] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] self.force_reraise() [ 693.429885] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 693.429885] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] raise self.value [ 693.429885] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 693.429885] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] updated_port = self._update_port( [ 693.429885] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 693.429885] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] _ensure_no_port_binding_failure(port) [ 693.429885] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 693.429885] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] raise exception.PortBindingFailed(port_id=port['id']) [ 693.429885] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] nova.exception.PortBindingFailed: Binding failed for port 1da29949-9732-45b3-a088-7b379ed73ad5, please check neutron logs for more information. [ 693.429885] env[62965]: ERROR nova.compute.manager [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] [ 693.430250] env[62965]: DEBUG nova.compute.utils [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Binding failed for port 1da29949-9732-45b3-a088-7b379ed73ad5, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 693.430948] env[62965]: DEBUG oslo_concurrency.lockutils [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.621s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.434987] env[62965]: DEBUG nova.compute.manager [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Build of instance 971f72f8-2686-4ced-a2a1-b7ab8a1ea024 was re-scheduled: Binding failed for port 1da29949-9732-45b3-a088-7b379ed73ad5, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 693.435396] env[62965]: DEBUG nova.compute.manager [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 693.435632] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Acquiring lock "refresh_cache-971f72f8-2686-4ced-a2a1-b7ab8a1ea024" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.435779] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Acquired lock "refresh_cache-971f72f8-2686-4ced-a2a1-b7ab8a1ea024" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.435938] env[62965]: DEBUG nova.network.neutron [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 693.641378] env[62965]: INFO nova.compute.manager [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] [instance: 8ac4264d-7d75-4c38-b8da-e3d90ffe9363] Took 1.03 seconds to deallocate network for instance. [ 693.681066] env[62965]: DEBUG oslo_vmware.api [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389767, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.475713} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.681327] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] 02f113d2-f991-4268-b9a4-59e2e6bcf7a6/02f113d2-f991-4268-b9a4-59e2e6bcf7a6.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 693.681529] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 693.681767] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4d0dabd2-1cea-4307-9b16-4a4e4c6c565b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.689413] env[62965]: DEBUG oslo_vmware.api [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Waiting for the task: (returnval){ [ 693.689413] env[62965]: value = "task-1389768" [ 693.689413] env[62965]: _type = "Task" [ 693.689413] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.697779] env[62965]: DEBUG oslo_vmware.api [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389768, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.956225] env[62965]: DEBUG nova.network.neutron [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 694.038704] env[62965]: DEBUG nova.network.neutron [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.200151] env[62965]: DEBUG oslo_vmware.api [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389768, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068417} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.202617] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 694.203540] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c00029b5-5b75-4bf1-adb0-0f9b5c2e1541 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.223220] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Reconfiguring VM instance instance-0000001f to attach disk [datastore1] 02f113d2-f991-4268-b9a4-59e2e6bcf7a6/02f113d2-f991-4268-b9a4-59e2e6bcf7a6.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 694.225631] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-424748ab-bc6b-4ee4-8161-e24e73ea11b9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.245358] env[62965]: DEBUG oslo_vmware.api [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Waiting for the task: (returnval){ [ 694.245358] env[62965]: value = "task-1389769" [ 694.245358] env[62965]: _type = "Task" [ 694.245358] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.255753] env[62965]: DEBUG oslo_vmware.api [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389769, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.320350] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ded22b7e-6d07-4ccf-8367-b665b90fbc32 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.329041] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28dd3c94-ad91-4331-8530-113453838e76 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.361223] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9436b97e-810d-4fe5-a942-af843abb0f59 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.369692] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-516c11ea-f04e-4f6a-a984-c7f9780f5040 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.383414] env[62965]: DEBUG nova.compute.provider_tree [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 694.541308] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Releasing lock "refresh_cache-971f72f8-2686-4ced-a2a1-b7ab8a1ea024" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.541415] env[62965]: DEBUG nova.compute.manager [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 694.541683] env[62965]: DEBUG nova.compute.manager [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 694.541919] env[62965]: DEBUG nova.network.neutron [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 694.558089] env[62965]: DEBUG nova.network.neutron [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 694.671872] env[62965]: INFO nova.scheduler.client.report [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] Deleted allocations for instance 8ac4264d-7d75-4c38-b8da-e3d90ffe9363 [ 694.756421] env[62965]: DEBUG oslo_vmware.api [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389769, 'name': ReconfigVM_Task, 'duration_secs': 0.287436} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.757118] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Reconfigured VM instance instance-0000001f to attach disk [datastore1] 02f113d2-f991-4268-b9a4-59e2e6bcf7a6/02f113d2-f991-4268-b9a4-59e2e6bcf7a6.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 694.757773] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-008d8b49-cf45-4895-9eb4-8e038a28b124 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.764424] env[62965]: DEBUG oslo_vmware.api [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Waiting for the task: (returnval){ [ 694.764424] env[62965]: value = "task-1389770" [ 694.764424] env[62965]: _type = "Task" [ 694.764424] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.772813] env[62965]: DEBUG oslo_vmware.api [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389770, 'name': Rename_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.886667] env[62965]: DEBUG nova.scheduler.client.report [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 695.060601] env[62965]: DEBUG nova.network.neutron [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.179752] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f38bdb2c-9565-46a0-b784-ce06dae728f0 tempest-InstanceActionsV221TestJSON-15046983 tempest-InstanceActionsV221TestJSON-15046983-project-member] Lock "8ac4264d-7d75-4c38-b8da-e3d90ffe9363" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.234s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.274530] env[62965]: DEBUG oslo_vmware.api [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389770, 'name': Rename_Task, 'duration_secs': 0.132762} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.274848] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 695.275109] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-28fea8fd-d73e-4c14-826d-7ca4c0ca6f0d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.285807] env[62965]: DEBUG oslo_vmware.api [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Waiting for the task: (returnval){ [ 695.285807] env[62965]: value = "task-1389771" [ 695.285807] env[62965]: _type = "Task" [ 695.285807] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.294282] env[62965]: DEBUG oslo_vmware.api [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389771, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.393923] env[62965]: DEBUG oslo_concurrency.lockutils [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.961s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.393923] env[62965]: ERROR nova.compute.manager [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 09855567-2998-47e2-9b3c-5e2f7b42f71f, please check neutron logs for more information. [ 695.393923] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Traceback (most recent call last): [ 695.393923] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 695.393923] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] self.driver.spawn(context, instance, image_meta, [ 695.393923] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 695.393923] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] self._vmops.spawn(context, instance, image_meta, injected_files, [ 695.393923] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 695.393923] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] vm_ref = self.build_virtual_machine(instance, [ 695.395111] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 695.395111] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] vif_infos = vmwarevif.get_vif_info(self._session, [ 695.395111] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 695.395111] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] for vif in network_info: [ 695.395111] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 695.395111] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] return self._sync_wrapper(fn, *args, **kwargs) [ 695.395111] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 695.395111] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] self.wait() [ 695.395111] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 695.395111] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] self[:] = self._gt.wait() [ 695.395111] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 695.395111] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] return self._exit_event.wait() [ 695.395111] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 695.395473] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] result = hub.switch() [ 695.395473] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 695.395473] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] return self.greenlet.switch() [ 695.395473] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 695.395473] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] result = function(*args, **kwargs) [ 695.395473] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 695.395473] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] return func(*args, **kwargs) [ 695.395473] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 695.395473] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] raise e [ 695.395473] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 695.395473] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] nwinfo = self.network_api.allocate_for_instance( [ 695.395473] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 695.395473] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] created_port_ids = self._update_ports_for_instance( [ 695.395794] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 695.395794] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] with excutils.save_and_reraise_exception(): [ 695.395794] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.395794] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] self.force_reraise() [ 695.395794] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.395794] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] raise self.value [ 695.395794] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 695.395794] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] updated_port = self._update_port( [ 695.395794] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.395794] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] _ensure_no_port_binding_failure(port) [ 695.395794] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.395794] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] raise exception.PortBindingFailed(port_id=port['id']) [ 695.396238] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] nova.exception.PortBindingFailed: Binding failed for port 09855567-2998-47e2-9b3c-5e2f7b42f71f, please check neutron logs for more information. [ 695.396238] env[62965]: ERROR nova.compute.manager [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] [ 695.396238] env[62965]: DEBUG nova.compute.utils [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Binding failed for port 09855567-2998-47e2-9b3c-5e2f7b42f71f, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 695.396238] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.162s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.397166] env[62965]: DEBUG nova.compute.manager [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Build of instance eaf9ca94-4f45-403e-bb06-e9a9adf58f38 was re-scheduled: Binding failed for port 09855567-2998-47e2-9b3c-5e2f7b42f71f, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 695.397590] env[62965]: DEBUG nova.compute.manager [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 695.397809] env[62965]: DEBUG oslo_concurrency.lockutils [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] Acquiring lock "refresh_cache-eaf9ca94-4f45-403e-bb06-e9a9adf58f38" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.397954] env[62965]: DEBUG oslo_concurrency.lockutils [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] Acquired lock "refresh_cache-eaf9ca94-4f45-403e-bb06-e9a9adf58f38" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.398125] env[62965]: DEBUG nova.network.neutron [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 695.564090] env[62965]: INFO nova.compute.manager [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 971f72f8-2686-4ced-a2a1-b7ab8a1ea024] Took 1.02 seconds to deallocate network for instance. [ 695.683150] env[62965]: DEBUG nova.compute.manager [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 695.799126] env[62965]: DEBUG oslo_vmware.api [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389771, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.923293] env[62965]: DEBUG nova.network.neutron [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 696.063234] env[62965]: DEBUG nova.network.neutron [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.204013] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.296684] env[62965]: DEBUG oslo_vmware.api [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389771, 'name': PowerOnVM_Task, 'duration_secs': 0.976127} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.296960] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 696.301068] env[62965]: DEBUG nova.compute.manager [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 696.301068] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-592f5e89-7271-4365-a64f-f3edde6c3792 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.354965] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daae5ae7-06aa-431a-82a9-a9a7a2f8fb69 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.370129] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1af1364c-8ee4-419c-bc12-3929bb44794d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.409298] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7140536b-3432-4d1a-a34e-2a43e2f830ef {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.419554] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07b9ee9f-4d00-4f61-8698-64f7d21a361e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.433556] env[62965]: DEBUG nova.compute.provider_tree [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 696.566499] env[62965]: DEBUG oslo_concurrency.lockutils [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] Releasing lock "refresh_cache-eaf9ca94-4f45-403e-bb06-e9a9adf58f38" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.566679] env[62965]: DEBUG nova.compute.manager [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 696.566891] env[62965]: DEBUG nova.compute.manager [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 696.567110] env[62965]: DEBUG nova.network.neutron [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 696.583465] env[62965]: DEBUG nova.network.neutron [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 696.611060] env[62965]: INFO nova.scheduler.client.report [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Deleted allocations for instance 971f72f8-2686-4ced-a2a1-b7ab8a1ea024 [ 696.822414] env[62965]: DEBUG oslo_concurrency.lockutils [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.937438] env[62965]: DEBUG nova.scheduler.client.report [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 697.088643] env[62965]: DEBUG nova.network.neutron [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.123565] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ff5f1043-bd6b-4842-b872-ecd424195d07 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Lock "971f72f8-2686-4ced-a2a1-b7ab8a1ea024" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 129.928s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 697.327167] env[62965]: INFO nova.compute.manager [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Rebuilding instance [ 697.374603] env[62965]: DEBUG nova.compute.manager [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 697.375596] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-002a3e4b-9df7-491e-8d47-c64faf9294c3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.443082] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.048s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 697.443476] env[62965]: ERROR nova.compute.manager [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 76d8df8a-8e64-4dd2-b6bd-f1726c488d7f, please check neutron logs for more information. [ 697.443476] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Traceback (most recent call last): [ 697.443476] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 697.443476] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] self.driver.spawn(context, instance, image_meta, [ 697.443476] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 697.443476] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] self._vmops.spawn(context, instance, image_meta, injected_files, [ 697.443476] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 697.443476] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] vm_ref = self.build_virtual_machine(instance, [ 697.443476] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 697.443476] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] vif_infos = vmwarevif.get_vif_info(self._session, [ 697.443476] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 697.443876] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] for vif in network_info: [ 697.443876] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 697.443876] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] return self._sync_wrapper(fn, *args, **kwargs) [ 697.443876] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 697.443876] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] self.wait() [ 697.443876] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 697.443876] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] self[:] = self._gt.wait() [ 697.443876] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 697.443876] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] return self._exit_event.wait() [ 697.443876] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 697.443876] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] result = hub.switch() [ 697.443876] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 697.443876] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] return self.greenlet.switch() [ 697.444250] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 697.444250] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] result = function(*args, **kwargs) [ 697.444250] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 697.444250] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] return func(*args, **kwargs) [ 697.444250] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 697.444250] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] raise e [ 697.444250] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 697.444250] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] nwinfo = self.network_api.allocate_for_instance( [ 697.444250] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 697.444250] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] created_port_ids = self._update_ports_for_instance( [ 697.444250] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 697.444250] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] with excutils.save_and_reraise_exception(): [ 697.444250] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 697.444589] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] self.force_reraise() [ 697.444589] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 697.444589] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] raise self.value [ 697.444589] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 697.444589] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] updated_port = self._update_port( [ 697.444589] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 697.444589] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] _ensure_no_port_binding_failure(port) [ 697.444589] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 697.444589] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] raise exception.PortBindingFailed(port_id=port['id']) [ 697.444589] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] nova.exception.PortBindingFailed: Binding failed for port 76d8df8a-8e64-4dd2-b6bd-f1726c488d7f, please check neutron logs for more information. [ 697.444589] env[62965]: ERROR nova.compute.manager [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] [ 697.444902] env[62965]: DEBUG nova.compute.utils [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Binding failed for port 76d8df8a-8e64-4dd2-b6bd-f1726c488d7f, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 697.445891] env[62965]: DEBUG nova.compute.manager [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Build of instance 79a4d3c8-ead6-4893-b0a7-89a8ff435e54 was re-scheduled: Binding failed for port 76d8df8a-8e64-4dd2-b6bd-f1726c488d7f, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 697.446361] env[62965]: DEBUG nova.compute.manager [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 697.446615] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Acquiring lock "refresh_cache-79a4d3c8-ead6-4893-b0a7-89a8ff435e54" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 697.446767] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Acquired lock "refresh_cache-79a4d3c8-ead6-4893-b0a7-89a8ff435e54" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.446926] env[62965]: DEBUG nova.network.neutron [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 697.447872] env[62965]: DEBUG oslo_concurrency.lockutils [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.464s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 697.591734] env[62965]: INFO nova.compute.manager [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] [instance: eaf9ca94-4f45-403e-bb06-e9a9adf58f38] Took 1.02 seconds to deallocate network for instance. [ 697.626099] env[62965]: DEBUG nova.compute.manager [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 697.971356] env[62965]: DEBUG nova.network.neutron [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 698.074324] env[62965]: DEBUG nova.network.neutron [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.156612] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 698.355177] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1e75824-0c60-40b1-8f93-2a7a43d27c6a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.364465] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7af89ffe-721a-49ec-9681-170a1a76af58 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.396505] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 698.396980] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9715ce08-801c-489d-9431-b54792de759f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.399150] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69f28125-08b7-4811-8b0a-17bd0bebb247 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.407867] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bb245b8-9471-452e-9fe9-eac2c5afca67 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.413099] env[62965]: DEBUG oslo_vmware.api [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Waiting for the task: (returnval){ [ 698.413099] env[62965]: value = "task-1389772" [ 698.413099] env[62965]: _type = "Task" [ 698.413099] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.425164] env[62965]: DEBUG nova.compute.provider_tree [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 698.432612] env[62965]: DEBUG oslo_vmware.api [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Task: {'id': task-1389772, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.577201] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Releasing lock "refresh_cache-79a4d3c8-ead6-4893-b0a7-89a8ff435e54" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 698.577435] env[62965]: DEBUG nova.compute.manager [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 698.577624] env[62965]: DEBUG nova.compute.manager [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 698.577790] env[62965]: DEBUG nova.network.neutron [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 698.597234] env[62965]: DEBUG nova.network.neutron [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 698.622118] env[62965]: INFO nova.scheduler.client.report [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] Deleted allocations for instance eaf9ca94-4f45-403e-bb06-e9a9adf58f38 [ 698.924884] env[62965]: DEBUG oslo_vmware.api [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Task: {'id': task-1389772, 'name': PowerOffVM_Task, 'duration_secs': 0.178681} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.925222] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 698.928016] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 698.928016] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de772d24-f3a4-4bbe-ac8d-60aead3ae379 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.930475] env[62965]: DEBUG nova.scheduler.client.report [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 698.941075] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 698.942575] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2a50d10f-1265-4d6f-84bb-edae0f4b9a4f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.977108] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 698.977108] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Deleting contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 698.977108] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Deleting the datastore file [datastore1] 02f113d2-f991-4268-b9a4-59e2e6bcf7a6 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 698.977108] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7e3c15a4-9a92-4d5a-a27f-3fb1a221737d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.983639] env[62965]: DEBUG oslo_vmware.api [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Waiting for the task: (returnval){ [ 698.983639] env[62965]: value = "task-1389774" [ 698.983639] env[62965]: _type = "Task" [ 698.983639] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.995598] env[62965]: DEBUG oslo_vmware.api [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Task: {'id': task-1389774, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.102192] env[62965]: DEBUG nova.network.neutron [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.131929] env[62965]: DEBUG oslo_concurrency.lockutils [None req-93bfe9ea-c8d7-470d-b66a-cab6a76a7fa6 tempest-ServerMetadataTestJSON-1143213976 tempest-ServerMetadataTestJSON-1143213976-project-member] Lock "eaf9ca94-4f45-403e-bb06-e9a9adf58f38" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.099s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.436455] env[62965]: DEBUG oslo_concurrency.lockutils [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.988s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.439020] env[62965]: ERROR nova.compute.manager [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 85c8faec-56d2-4fad-86ff-cc4710a93278, please check neutron logs for more information. [ 699.439020] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Traceback (most recent call last): [ 699.439020] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 699.439020] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] self.driver.spawn(context, instance, image_meta, [ 699.439020] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 699.439020] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 699.439020] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 699.439020] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] vm_ref = self.build_virtual_machine(instance, [ 699.439020] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 699.439020] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] vif_infos = vmwarevif.get_vif_info(self._session, [ 699.439020] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 699.439427] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] for vif in network_info: [ 699.439427] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 699.439427] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] return self._sync_wrapper(fn, *args, **kwargs) [ 699.439427] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 699.439427] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] self.wait() [ 699.439427] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 699.439427] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] self[:] = self._gt.wait() [ 699.439427] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 699.439427] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] return self._exit_event.wait() [ 699.439427] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 699.439427] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] current.throw(*self._exc) [ 699.439427] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 699.439427] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] result = function(*args, **kwargs) [ 699.439776] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 699.439776] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] return func(*args, **kwargs) [ 699.439776] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 699.439776] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] raise e [ 699.439776] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 699.439776] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] nwinfo = self.network_api.allocate_for_instance( [ 699.439776] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 699.439776] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] created_port_ids = self._update_ports_for_instance( [ 699.439776] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 699.439776] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] with excutils.save_and_reraise_exception(): [ 699.439776] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.439776] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] self.force_reraise() [ 699.439776] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.440150] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] raise self.value [ 699.440150] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 699.440150] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] updated_port = self._update_port( [ 699.440150] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.440150] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] _ensure_no_port_binding_failure(port) [ 699.440150] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.440150] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] raise exception.PortBindingFailed(port_id=port['id']) [ 699.440150] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] nova.exception.PortBindingFailed: Binding failed for port 85c8faec-56d2-4fad-86ff-cc4710a93278, please check neutron logs for more information. [ 699.440150] env[62965]: ERROR nova.compute.manager [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] [ 699.440150] env[62965]: DEBUG nova.compute.utils [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Binding failed for port 85c8faec-56d2-4fad-86ff-cc4710a93278, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 699.442246] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.236s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 699.444015] env[62965]: INFO nova.compute.claims [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 699.448085] env[62965]: DEBUG nova.compute.manager [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Build of instance 7eee167f-50e3-48d5-9c6d-9115df9bb0b9 was re-scheduled: Binding failed for port 85c8faec-56d2-4fad-86ff-cc4710a93278, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 699.448085] env[62965]: DEBUG nova.compute.manager [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 699.448085] env[62965]: DEBUG oslo_concurrency.lockutils [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Acquiring lock "refresh_cache-7eee167f-50e3-48d5-9c6d-9115df9bb0b9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.448085] env[62965]: DEBUG oslo_concurrency.lockutils [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Acquired lock "refresh_cache-7eee167f-50e3-48d5-9c6d-9115df9bb0b9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.448287] env[62965]: DEBUG nova.network.neutron [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 699.495190] env[62965]: DEBUG oslo_vmware.api [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Task: {'id': task-1389774, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.100753} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.495831] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 699.496164] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Deleted contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 699.497377] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 699.607261] env[62965]: INFO nova.compute.manager [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] [instance: 79a4d3c8-ead6-4893-b0a7-89a8ff435e54] Took 1.03 seconds to deallocate network for instance. [ 699.634485] env[62965]: DEBUG nova.compute.manager [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 699.983228] env[62965]: DEBUG nova.network.neutron [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 700.091174] env[62965]: DEBUG nova.network.neutron [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.172524] env[62965]: DEBUG oslo_concurrency.lockutils [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 700.540949] env[62965]: DEBUG nova.virt.hardware [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 700.543134] env[62965]: DEBUG nova.virt.hardware [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 700.543314] env[62965]: DEBUG nova.virt.hardware [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 700.543515] env[62965]: DEBUG nova.virt.hardware [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 700.543718] env[62965]: DEBUG nova.virt.hardware [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 700.543821] env[62965]: DEBUG nova.virt.hardware [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 700.544043] env[62965]: DEBUG nova.virt.hardware [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 700.544205] env[62965]: DEBUG nova.virt.hardware [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 700.544386] env[62965]: DEBUG nova.virt.hardware [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 700.545014] env[62965]: DEBUG nova.virt.hardware [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 700.545014] env[62965]: DEBUG nova.virt.hardware [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 700.545648] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8cc9000-9b63-4887-b7fc-641dcad66d6b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.558585] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3519ee04-990c-43b7-9ea9-14b48510841c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.575930] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Instance VIF info [] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 700.581913] env[62965]: DEBUG oslo.service.loopingcall [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 700.585863] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 700.585863] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3f3f4446-7cb4-48eb-a6d9-a5638640a704 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.600396] env[62965]: DEBUG oslo_concurrency.lockutils [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Releasing lock "refresh_cache-7eee167f-50e3-48d5-9c6d-9115df9bb0b9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.601975] env[62965]: DEBUG nova.compute.manager [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 700.601975] env[62965]: DEBUG nova.compute.manager [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 700.601975] env[62965]: DEBUG nova.network.neutron [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 700.613167] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 700.613167] env[62965]: value = "task-1389775" [ 700.613167] env[62965]: _type = "Task" [ 700.613167] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.632105] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389775, 'name': CreateVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.646036] env[62965]: DEBUG nova.network.neutron [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 700.646646] env[62965]: INFO nova.scheduler.client.report [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Deleted allocations for instance 79a4d3c8-ead6-4893-b0a7-89a8ff435e54 [ 700.918716] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d2d9c96-0155-4e49-9284-aa9f623bd995 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.927086] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99a3611e-c260-4595-84db-034a3a5d8895 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.961109] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-044b9ac4-1de4-4acd-bab9-edd39401bed1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.969522] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8ce0b0d-ba57-46ee-8528-90390901ea5c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.986834] env[62965]: DEBUG nova.compute.provider_tree [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 701.125215] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389775, 'name': CreateVM_Task, 'duration_secs': 0.271831} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.125460] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 701.126593] env[62965]: DEBUG oslo_concurrency.lockutils [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.126905] env[62965]: DEBUG oslo_concurrency.lockutils [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.127563] env[62965]: DEBUG oslo_concurrency.lockutils [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Acquired external semaphore "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 701.130026] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb539508-f01b-4e6b-ad08-5295c1771392 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.138170] env[62965]: DEBUG oslo_vmware.api [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Waiting for the task: (returnval){ [ 701.138170] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]522d92db-0102-e399-c89a-ae41e5eeeb67" [ 701.138170] env[62965]: _type = "Task" [ 701.138170] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.147984] env[62965]: DEBUG oslo_vmware.api [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]522d92db-0102-e399-c89a-ae41e5eeeb67, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.154508] env[62965]: DEBUG nova.network.neutron [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.158441] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c129206b-2423-4d34-8547-3335f6ed6db6 tempest-ListServerFiltersTestJSON-1329335703 tempest-ListServerFiltersTestJSON-1329335703-project-member] Lock "79a4d3c8-ead6-4893-b0a7-89a8ff435e54" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 132.123s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.487688] env[62965]: DEBUG nova.scheduler.client.report [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 701.649827] env[62965]: DEBUG oslo_vmware.api [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]522d92db-0102-e399-c89a-ae41e5eeeb67, 'name': SearchDatastore_Task, 'duration_secs': 0.010752} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.649827] env[62965]: DEBUG oslo_concurrency.lockutils [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.652254] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 701.652254] env[62965]: DEBUG oslo_concurrency.lockutils [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.652254] env[62965]: DEBUG oslo_concurrency.lockutils [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.652254] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 701.652254] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ca7ae599-2e96-4616-8a5f-04a0e4466db5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.661021] env[62965]: INFO nova.compute.manager [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] [instance: 7eee167f-50e3-48d5-9c6d-9115df9bb0b9] Took 1.06 seconds to deallocate network for instance. [ 701.661021] env[62965]: DEBUG nova.compute.manager [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 701.666020] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 701.666420] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Folder [datastore1] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 701.667595] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c624ce9-5eee-4459-9495-dc654e2e062e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.674292] env[62965]: DEBUG oslo_vmware.api [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Waiting for the task: (returnval){ [ 701.674292] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52b2e9ec-d282-4290-cddf-91543dc00aa3" [ 701.674292] env[62965]: _type = "Task" [ 701.674292] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.686527] env[62965]: DEBUG oslo_vmware.api [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52b2e9ec-d282-4290-cddf-91543dc00aa3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.722806] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] Acquiring lock "a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 701.722806] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] Lock "a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.991823] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.551s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.992370] env[62965]: DEBUG nova.compute.manager [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 701.998414] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.916s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 702.186984] env[62965]: DEBUG oslo_vmware.api [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52b2e9ec-d282-4290-cddf-91543dc00aa3, 'name': SearchDatastore_Task, 'duration_secs': 0.009641} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.187843] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c946fb12-6da8-41a8-a209-6ca190abb707 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.195082] env[62965]: DEBUG oslo_vmware.api [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Waiting for the task: (returnval){ [ 702.195082] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5285878c-a83f-44d5-b419-8ed52829cf5e" [ 702.195082] env[62965]: _type = "Task" [ 702.195082] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.199550] env[62965]: DEBUG oslo_concurrency.lockutils [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.207269] env[62965]: DEBUG oslo_vmware.api [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5285878c-a83f-44d5-b419-8ed52829cf5e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.502396] env[62965]: DEBUG nova.compute.utils [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 702.506720] env[62965]: DEBUG nova.compute.manager [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 702.506889] env[62965]: DEBUG nova.network.neutron [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 702.566229] env[62965]: DEBUG nova.policy [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '796be5c70f074929943bc1196791eba3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '07cc096d8b1e4f8284d9349b038e430e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 702.691837] env[62965]: INFO nova.scheduler.client.report [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Deleted allocations for instance 7eee167f-50e3-48d5-9c6d-9115df9bb0b9 [ 702.712501] env[62965]: DEBUG oslo_vmware.api [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5285878c-a83f-44d5-b419-8ed52829cf5e, 'name': SearchDatastore_Task, 'duration_secs': 0.009582} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.712501] env[62965]: DEBUG oslo_concurrency.lockutils [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 702.712501] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Copying Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] 02f113d2-f991-4268-b9a4-59e2e6bcf7a6/02f113d2-f991-4268-b9a4-59e2e6bcf7a6.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 702.712501] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8c972313-74c6-4027-8178-2f6e1fe063d8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.719820] env[62965]: DEBUG oslo_vmware.api [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Waiting for the task: (returnval){ [ 702.719820] env[62965]: value = "task-1389776" [ 702.719820] env[62965]: _type = "Task" [ 702.719820] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.735179] env[62965]: DEBUG oslo_vmware.api [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Task: {'id': task-1389776, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.940951] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-555ae10c-bc6f-4aa8-aa04-e59fb651f115 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.953043] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b72dc15-f48a-4621-bb28-a83970889dc5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.997936] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7b0f893-924b-43f1-8f62-14315a4f9a20 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.012162] env[62965]: DEBUG nova.compute.manager [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 703.016297] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a9a958c-c9db-45b9-a0fa-24060d8ccfee {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.034729] env[62965]: DEBUG nova.compute.provider_tree [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 703.078794] env[62965]: DEBUG nova.network.neutron [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Successfully created port: 42ad3568-1e88-47f0-bf93-9f3dfe2266c8 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 703.212588] env[62965]: DEBUG oslo_concurrency.lockutils [None req-cc5c52eb-741b-4002-a3eb-2cc38c55d1a1 tempest-DeleteServersAdminTestJSON-1402977290 tempest-DeleteServersAdminTestJSON-1402977290-project-member] Lock "7eee167f-50e3-48d5-9c6d-9115df9bb0b9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 130.975s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.230153] env[62965]: DEBUG oslo_vmware.api [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Task: {'id': task-1389776, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.492542} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.230538] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Copied Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] 02f113d2-f991-4268-b9a4-59e2e6bcf7a6/02f113d2-f991-4268-b9a4-59e2e6bcf7a6.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 703.230661] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 703.230954] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7a3dbd20-e35c-4442-a34d-8beee3203fd0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.239653] env[62965]: DEBUG oslo_vmware.api [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Waiting for the task: (returnval){ [ 703.239653] env[62965]: value = "task-1389777" [ 703.239653] env[62965]: _type = "Task" [ 703.239653] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.250786] env[62965]: DEBUG oslo_vmware.api [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Task: {'id': task-1389777, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.540229] env[62965]: DEBUG nova.scheduler.client.report [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 703.715271] env[62965]: DEBUG nova.compute.manager [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 703.751941] env[62965]: DEBUG oslo_vmware.api [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Task: {'id': task-1389777, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062788} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.753592] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 703.756026] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32554005-bde9-4d00-b9b0-ef779973afa3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.776531] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Reconfiguring VM instance instance-0000001f to attach disk [datastore1] 02f113d2-f991-4268-b9a4-59e2e6bcf7a6/02f113d2-f991-4268-b9a4-59e2e6bcf7a6.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 703.777360] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4b3a8cfd-a11a-4311-b079-10f5c4480dd6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.806350] env[62965]: DEBUG oslo_vmware.api [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Waiting for the task: (returnval){ [ 703.806350] env[62965]: value = "task-1389778" [ 703.806350] env[62965]: _type = "Task" [ 703.806350] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.817921] env[62965]: DEBUG oslo_vmware.api [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Task: {'id': task-1389778, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.929125] env[62965]: DEBUG nova.compute.manager [req-60605412-eb5c-4c9d-9efb-ef7f3a790c6d req-00bf73c4-b022-406c-bd80-4fc3603a5c8b service nova] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Received event network-changed-42ad3568-1e88-47f0-bf93-9f3dfe2266c8 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 703.929125] env[62965]: DEBUG nova.compute.manager [req-60605412-eb5c-4c9d-9efb-ef7f3a790c6d req-00bf73c4-b022-406c-bd80-4fc3603a5c8b service nova] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Refreshing instance network info cache due to event network-changed-42ad3568-1e88-47f0-bf93-9f3dfe2266c8. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 703.929125] env[62965]: DEBUG oslo_concurrency.lockutils [req-60605412-eb5c-4c9d-9efb-ef7f3a790c6d req-00bf73c4-b022-406c-bd80-4fc3603a5c8b service nova] Acquiring lock "refresh_cache-4519d37a-f8e6-4a25-99a1-399925199a58" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 703.929719] env[62965]: DEBUG oslo_concurrency.lockutils [req-60605412-eb5c-4c9d-9efb-ef7f3a790c6d req-00bf73c4-b022-406c-bd80-4fc3603a5c8b service nova] Acquired lock "refresh_cache-4519d37a-f8e6-4a25-99a1-399925199a58" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.931122] env[62965]: DEBUG nova.network.neutron [req-60605412-eb5c-4c9d-9efb-ef7f3a790c6d req-00bf73c4-b022-406c-bd80-4fc3603a5c8b service nova] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Refreshing network info cache for port 42ad3568-1e88-47f0-bf93-9f3dfe2266c8 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 704.029918] env[62965]: DEBUG nova.compute.manager [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 704.047792] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.050s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 704.048480] env[62965]: ERROR nova.compute.manager [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port dacdb675-46d2-428a-9408-131ed8d7743b, please check neutron logs for more information. [ 704.048480] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Traceback (most recent call last): [ 704.048480] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 704.048480] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] self.driver.spawn(context, instance, image_meta, [ 704.048480] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 704.048480] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] self._vmops.spawn(context, instance, image_meta, injected_files, [ 704.048480] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 704.048480] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] vm_ref = self.build_virtual_machine(instance, [ 704.048480] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 704.048480] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] vif_infos = vmwarevif.get_vif_info(self._session, [ 704.048480] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 704.048819] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] for vif in network_info: [ 704.048819] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 704.048819] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] return self._sync_wrapper(fn, *args, **kwargs) [ 704.048819] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 704.048819] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] self.wait() [ 704.048819] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 704.048819] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] self[:] = self._gt.wait() [ 704.048819] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 704.048819] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] return self._exit_event.wait() [ 704.048819] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 704.048819] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] current.throw(*self._exc) [ 704.048819] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 704.048819] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] result = function(*args, **kwargs) [ 704.049160] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 704.049160] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] return func(*args, **kwargs) [ 704.049160] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 704.049160] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] raise e [ 704.049160] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 704.049160] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] nwinfo = self.network_api.allocate_for_instance( [ 704.049160] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 704.049160] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] created_port_ids = self._update_ports_for_instance( [ 704.049160] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 704.049160] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] with excutils.save_and_reraise_exception(): [ 704.049160] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 704.049160] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] self.force_reraise() [ 704.049160] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 704.049542] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] raise self.value [ 704.049542] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 704.049542] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] updated_port = self._update_port( [ 704.049542] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 704.049542] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] _ensure_no_port_binding_failure(port) [ 704.049542] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 704.049542] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] raise exception.PortBindingFailed(port_id=port['id']) [ 704.049542] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] nova.exception.PortBindingFailed: Binding failed for port dacdb675-46d2-428a-9408-131ed8d7743b, please check neutron logs for more information. [ 704.049542] env[62965]: ERROR nova.compute.manager [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] [ 704.049542] env[62965]: DEBUG nova.compute.utils [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Binding failed for port dacdb675-46d2-428a-9408-131ed8d7743b, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 704.056451] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.251s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.058497] env[62965]: INFO nova.compute.claims [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 704.064562] env[62965]: DEBUG nova.compute.manager [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Build of instance f45f605f-cb91-4169-879c-bf9e6ce9c989 was re-scheduled: Binding failed for port dacdb675-46d2-428a-9408-131ed8d7743b, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 704.064562] env[62965]: DEBUG nova.compute.manager [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 704.064562] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] Acquiring lock "refresh_cache-f45f605f-cb91-4169-879c-bf9e6ce9c989" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.064562] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] Acquired lock "refresh_cache-f45f605f-cb91-4169-879c-bf9e6ce9c989" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.064903] env[62965]: DEBUG nova.network.neutron [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 704.075384] env[62965]: DEBUG nova.virt.hardware [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 704.075621] env[62965]: DEBUG nova.virt.hardware [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 704.075774] env[62965]: DEBUG nova.virt.hardware [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 704.075948] env[62965]: DEBUG nova.virt.hardware [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 704.076111] env[62965]: DEBUG nova.virt.hardware [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 704.076469] env[62965]: DEBUG nova.virt.hardware [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 704.076469] env[62965]: DEBUG nova.virt.hardware [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 704.076613] env[62965]: DEBUG nova.virt.hardware [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 704.076775] env[62965]: DEBUG nova.virt.hardware [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 704.076928] env[62965]: DEBUG nova.virt.hardware [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 704.077107] env[62965]: DEBUG nova.virt.hardware [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 704.077984] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f2105c8-502f-4258-9eeb-1003918a9a99 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.089253] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d3dd04e-1b63-4613-89f3-55f620658e0e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.123846] env[62965]: ERROR nova.compute.manager [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 42ad3568-1e88-47f0-bf93-9f3dfe2266c8, please check neutron logs for more information. [ 704.123846] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 704.123846] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 704.123846] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 704.123846] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 704.123846] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 704.123846] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 704.123846] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 704.123846] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 704.123846] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 704.123846] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 704.123846] env[62965]: ERROR nova.compute.manager raise self.value [ 704.123846] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 704.123846] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 704.123846] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 704.123846] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 704.124406] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 704.124406] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 704.124406] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 42ad3568-1e88-47f0-bf93-9f3dfe2266c8, please check neutron logs for more information. [ 704.124406] env[62965]: ERROR nova.compute.manager [ 704.124406] env[62965]: Traceback (most recent call last): [ 704.124406] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 704.124406] env[62965]: listener.cb(fileno) [ 704.124406] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 704.124406] env[62965]: result = function(*args, **kwargs) [ 704.124406] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 704.124406] env[62965]: return func(*args, **kwargs) [ 704.124406] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 704.124406] env[62965]: raise e [ 704.124406] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 704.124406] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 704.124406] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 704.124406] env[62965]: created_port_ids = self._update_ports_for_instance( [ 704.124406] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 704.124406] env[62965]: with excutils.save_and_reraise_exception(): [ 704.124406] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 704.124406] env[62965]: self.force_reraise() [ 704.124406] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 704.124406] env[62965]: raise self.value [ 704.124406] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 704.124406] env[62965]: updated_port = self._update_port( [ 704.124406] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 704.124406] env[62965]: _ensure_no_port_binding_failure(port) [ 704.124406] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 704.124406] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 704.125328] env[62965]: nova.exception.PortBindingFailed: Binding failed for port 42ad3568-1e88-47f0-bf93-9f3dfe2266c8, please check neutron logs for more information. [ 704.125328] env[62965]: Removing descriptor: 19 [ 704.125328] env[62965]: ERROR nova.compute.manager [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 42ad3568-1e88-47f0-bf93-9f3dfe2266c8, please check neutron logs for more information. [ 704.125328] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Traceback (most recent call last): [ 704.125328] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 704.125328] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] yield resources [ 704.125328] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 704.125328] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] self.driver.spawn(context, instance, image_meta, [ 704.125328] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 704.125328] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] self._vmops.spawn(context, instance, image_meta, injected_files, [ 704.125328] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 704.125328] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] vm_ref = self.build_virtual_machine(instance, [ 704.125837] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 704.125837] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] vif_infos = vmwarevif.get_vif_info(self._session, [ 704.125837] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 704.125837] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] for vif in network_info: [ 704.125837] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 704.125837] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] return self._sync_wrapper(fn, *args, **kwargs) [ 704.125837] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 704.125837] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] self.wait() [ 704.125837] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 704.125837] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] self[:] = self._gt.wait() [ 704.125837] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 704.125837] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] return self._exit_event.wait() [ 704.125837] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 704.126237] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] result = hub.switch() [ 704.126237] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 704.126237] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] return self.greenlet.switch() [ 704.126237] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 704.126237] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] result = function(*args, **kwargs) [ 704.126237] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 704.126237] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] return func(*args, **kwargs) [ 704.126237] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 704.126237] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] raise e [ 704.126237] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 704.126237] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] nwinfo = self.network_api.allocate_for_instance( [ 704.126237] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 704.126237] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] created_port_ids = self._update_ports_for_instance( [ 704.126655] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 704.126655] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] with excutils.save_and_reraise_exception(): [ 704.126655] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 704.126655] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] self.force_reraise() [ 704.126655] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 704.126655] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] raise self.value [ 704.126655] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 704.126655] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] updated_port = self._update_port( [ 704.126655] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 704.126655] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] _ensure_no_port_binding_failure(port) [ 704.126655] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 704.126655] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] raise exception.PortBindingFailed(port_id=port['id']) [ 704.127024] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] nova.exception.PortBindingFailed: Binding failed for port 42ad3568-1e88-47f0-bf93-9f3dfe2266c8, please check neutron logs for more information. [ 704.127024] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] [ 704.127024] env[62965]: INFO nova.compute.manager [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Terminating instance [ 704.242399] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.317690] env[62965]: DEBUG oslo_vmware.api [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Task: {'id': task-1389778, 'name': ReconfigVM_Task, 'duration_secs': 0.317114} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.317995] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Reconfigured VM instance instance-0000001f to attach disk [datastore1] 02f113d2-f991-4268-b9a4-59e2e6bcf7a6/02f113d2-f991-4268-b9a4-59e2e6bcf7a6.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 704.318675] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a4e38cd1-e9ad-4a02-bd2b-492a15f9a99e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.326045] env[62965]: DEBUG oslo_vmware.api [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Waiting for the task: (returnval){ [ 704.326045] env[62965]: value = "task-1389779" [ 704.326045] env[62965]: _type = "Task" [ 704.326045] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.335413] env[62965]: DEBUG oslo_vmware.api [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Task: {'id': task-1389779, 'name': Rename_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.454540] env[62965]: DEBUG nova.network.neutron [req-60605412-eb5c-4c9d-9efb-ef7f3a790c6d req-00bf73c4-b022-406c-bd80-4fc3603a5c8b service nova] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 704.508125] env[62965]: DEBUG nova.network.neutron [req-60605412-eb5c-4c9d-9efb-ef7f3a790c6d req-00bf73c4-b022-406c-bd80-4fc3603a5c8b service nova] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.589583] env[62965]: DEBUG nova.network.neutron [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 704.632305] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Acquiring lock "refresh_cache-4519d37a-f8e6-4a25-99a1-399925199a58" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.734158] env[62965]: DEBUG nova.network.neutron [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.838163] env[62965]: DEBUG oslo_vmware.api [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Task: {'id': task-1389779, 'name': Rename_Task, 'duration_secs': 0.143593} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.838475] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 704.838734] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5fa18e2d-157b-47b6-9106-360328407234 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.845475] env[62965]: DEBUG oslo_vmware.api [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Waiting for the task: (returnval){ [ 704.845475] env[62965]: value = "task-1389780" [ 704.845475] env[62965]: _type = "Task" [ 704.845475] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.857329] env[62965]: DEBUG oslo_vmware.api [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Task: {'id': task-1389780, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.014588] env[62965]: DEBUG oslo_concurrency.lockutils [req-60605412-eb5c-4c9d-9efb-ef7f3a790c6d req-00bf73c4-b022-406c-bd80-4fc3603a5c8b service nova] Releasing lock "refresh_cache-4519d37a-f8e6-4a25-99a1-399925199a58" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 705.014588] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Acquired lock "refresh_cache-4519d37a-f8e6-4a25-99a1-399925199a58" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.014588] env[62965]: DEBUG nova.network.neutron [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 705.238472] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] Releasing lock "refresh_cache-f45f605f-cb91-4169-879c-bf9e6ce9c989" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 705.238472] env[62965]: DEBUG nova.compute.manager [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 705.238472] env[62965]: DEBUG nova.compute.manager [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 705.238472] env[62965]: DEBUG nova.network.neutron [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 705.263090] env[62965]: DEBUG nova.network.neutron [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 705.359615] env[62965]: DEBUG oslo_vmware.api [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Task: {'id': task-1389780, 'name': PowerOnVM_Task, 'duration_secs': 0.436871} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.359695] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 705.359836] env[62965]: DEBUG nova.compute.manager [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 705.360832] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd380db3-afe0-4f68-b96a-3ed06b024b38 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.538291] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4cf2a8c-846e-481b-bc93-27dee780a6d8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.544244] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab609848-c0d1-4465-8575-184fd82d1e6a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.550200] env[62965]: DEBUG nova.network.neutron [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 705.586479] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4d7ca74-c1ad-4fec-9581-ba91f17375a8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.595017] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2872cdb2-5448-408a-a96e-71151b982c10 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.610397] env[62965]: DEBUG nova.compute.provider_tree [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 705.687639] env[62965]: DEBUG nova.network.neutron [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.770404] env[62965]: DEBUG nova.network.neutron [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.887451] env[62965]: DEBUG oslo_concurrency.lockutils [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 705.964560] env[62965]: DEBUG nova.compute.manager [req-dd23468d-418c-4a2a-9243-cb2175a8152d req-51333ca8-320a-4ee6-9068-2dd07202c7c0 service nova] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Received event network-vif-deleted-42ad3568-1e88-47f0-bf93-9f3dfe2266c8 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 706.112901] env[62965]: DEBUG nova.scheduler.client.report [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 706.192018] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Releasing lock "refresh_cache-4519d37a-f8e6-4a25-99a1-399925199a58" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.192018] env[62965]: DEBUG nova.compute.manager [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 706.192018] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 706.192018] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6df2dd87-8353-47f8-a2ad-b037c64acb9d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.209594] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-319cacc2-8e87-4caa-b623-28c2345497d7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.236689] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4519d37a-f8e6-4a25-99a1-399925199a58 could not be found. [ 706.236916] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 706.237280] env[62965]: INFO nova.compute.manager [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Took 0.05 seconds to destroy the instance on the hypervisor. [ 706.237752] env[62965]: DEBUG oslo.service.loopingcall [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 706.240336] env[62965]: DEBUG nova.compute.manager [-] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 706.240336] env[62965]: DEBUG nova.network.neutron [-] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 706.261773] env[62965]: DEBUG nova.network.neutron [-] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 706.272076] env[62965]: INFO nova.compute.manager [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] [instance: f45f605f-cb91-4169-879c-bf9e6ce9c989] Took 1.03 seconds to deallocate network for instance. [ 706.618102] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.562s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.620431] env[62965]: DEBUG nova.compute.manager [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 706.624283] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.792s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.624283] env[62965]: INFO nova.compute.claims [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 706.770408] env[62965]: DEBUG nova.network.neutron [-] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.911132] env[62965]: DEBUG oslo_concurrency.lockutils [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Acquiring lock "02f113d2-f991-4268-b9a4-59e2e6bcf7a6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.911436] env[62965]: DEBUG oslo_concurrency.lockutils [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Lock "02f113d2-f991-4268-b9a4-59e2e6bcf7a6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.911647] env[62965]: DEBUG oslo_concurrency.lockutils [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Acquiring lock "02f113d2-f991-4268-b9a4-59e2e6bcf7a6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.911822] env[62965]: DEBUG oslo_concurrency.lockutils [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Lock "02f113d2-f991-4268-b9a4-59e2e6bcf7a6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.911979] env[62965]: DEBUG oslo_concurrency.lockutils [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Lock "02f113d2-f991-4268-b9a4-59e2e6bcf7a6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.914421] env[62965]: INFO nova.compute.manager [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Terminating instance [ 707.131227] env[62965]: DEBUG nova.compute.utils [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 707.136989] env[62965]: DEBUG nova.compute.manager [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 707.137266] env[62965]: DEBUG nova.network.neutron [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 707.225417] env[62965]: DEBUG nova.policy [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '600137dbd885485c9345045cc6c10fca', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b536b11c194a4acabf2a283d75485aed', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 707.273573] env[62965]: INFO nova.compute.manager [-] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Took 1.03 seconds to deallocate network for instance. [ 707.277703] env[62965]: DEBUG nova.compute.claims [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 707.277800] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.319920] env[62965]: INFO nova.scheduler.client.report [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] Deleted allocations for instance f45f605f-cb91-4169-879c-bf9e6ce9c989 [ 707.420895] env[62965]: DEBUG oslo_concurrency.lockutils [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Acquiring lock "refresh_cache-02f113d2-f991-4268-b9a4-59e2e6bcf7a6" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.421106] env[62965]: DEBUG oslo_concurrency.lockutils [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Acquired lock "refresh_cache-02f113d2-f991-4268-b9a4-59e2e6bcf7a6" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.421280] env[62965]: DEBUG nova.network.neutron [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 707.634454] env[62965]: DEBUG nova.compute.manager [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 707.718343] env[62965]: DEBUG nova.network.neutron [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Successfully created port: cfb0fc29-145a-45a9-8e88-4fb6176ed1e4 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 707.835311] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a5f0b609-52fb-4826-be8c-1ae90dbbdf1a tempest-ImagesNegativeTestJSON-694409788 tempest-ImagesNegativeTestJSON-694409788-project-member] Lock "f45f605f-cb91-4169-879c-bf9e6ce9c989" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 134.147s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.835575] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquiring lock "21bf78fb-4da7-4484-9678-b22a6d944c0b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.835786] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "21bf78fb-4da7-4484-9678-b22a6d944c0b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.947164] env[62965]: DEBUG nova.network.neutron [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 708.019082] env[62965]: DEBUG nova.network.neutron [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.136067] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fd08762-2070-4990-82e1-87d04835cdd9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.155104] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a63c112a-1c5b-4835-99f9-7adfb3003649 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.191252] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-168b93fb-69c1-4500-819e-e484685eb5da {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.207158] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6f15c53-ec29-46c1-8336-81b8c169c33a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.224551] env[62965]: DEBUG nova.compute.provider_tree [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 708.339525] env[62965]: DEBUG nova.compute.manager [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 708.520581] env[62965]: DEBUG oslo_concurrency.lockutils [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Releasing lock "refresh_cache-02f113d2-f991-4268-b9a4-59e2e6bcf7a6" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.521174] env[62965]: DEBUG nova.compute.manager [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 708.522053] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 708.522773] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6cc7b93-a5f3-4402-872b-53311275a23c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.531821] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 708.532296] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-093166d4-142f-411a-ad20-17263f753054 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.540362] env[62965]: DEBUG oslo_vmware.api [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Waiting for the task: (returnval){ [ 708.540362] env[62965]: value = "task-1389781" [ 708.540362] env[62965]: _type = "Task" [ 708.540362] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.547854] env[62965]: DEBUG oslo_vmware.api [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389781, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.649423] env[62965]: DEBUG nova.compute.manager [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 708.685912] env[62965]: DEBUG nova.virt.hardware [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 708.686195] env[62965]: DEBUG nova.virt.hardware [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 708.686444] env[62965]: DEBUG nova.virt.hardware [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 708.686683] env[62965]: DEBUG nova.virt.hardware [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 708.686882] env[62965]: DEBUG nova.virt.hardware [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 708.687099] env[62965]: DEBUG nova.virt.hardware [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 708.687367] env[62965]: DEBUG nova.virt.hardware [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 708.687590] env[62965]: DEBUG nova.virt.hardware [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 708.687851] env[62965]: DEBUG nova.virt.hardware [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 708.688089] env[62965]: DEBUG nova.virt.hardware [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 708.688319] env[62965]: DEBUG nova.virt.hardware [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 708.689046] env[62965]: DEBUG oslo_concurrency.lockutils [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] Acquiring lock "13d20f82-fa4f-41ab-b08f-5b64da7dddc0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 708.689325] env[62965]: DEBUG oslo_concurrency.lockutils [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] Lock "13d20f82-fa4f-41ab-b08f-5b64da7dddc0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 708.690184] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39683138-d572-4dfc-a2c3-4d36a721a741 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.698700] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b583f404-eaa2-4d7b-8e63-928a713213b1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.713867] env[62965]: ERROR nova.compute.manager [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port cfb0fc29-145a-45a9-8e88-4fb6176ed1e4, please check neutron logs for more information. [ 708.713867] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 708.713867] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 708.713867] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 708.713867] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 708.713867] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 708.713867] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 708.713867] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 708.713867] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 708.713867] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 708.713867] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 708.713867] env[62965]: ERROR nova.compute.manager raise self.value [ 708.713867] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 708.713867] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 708.713867] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 708.713867] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 708.714402] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 708.714402] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 708.714402] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port cfb0fc29-145a-45a9-8e88-4fb6176ed1e4, please check neutron logs for more information. [ 708.714402] env[62965]: ERROR nova.compute.manager [ 708.714402] env[62965]: Traceback (most recent call last): [ 708.714402] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 708.714402] env[62965]: listener.cb(fileno) [ 708.714402] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 708.714402] env[62965]: result = function(*args, **kwargs) [ 708.714402] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 708.714402] env[62965]: return func(*args, **kwargs) [ 708.714402] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 708.714402] env[62965]: raise e [ 708.714402] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 708.714402] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 708.714402] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 708.714402] env[62965]: created_port_ids = self._update_ports_for_instance( [ 708.714402] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 708.714402] env[62965]: with excutils.save_and_reraise_exception(): [ 708.714402] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 708.714402] env[62965]: self.force_reraise() [ 708.714402] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 708.714402] env[62965]: raise self.value [ 708.714402] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 708.714402] env[62965]: updated_port = self._update_port( [ 708.714402] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 708.714402] env[62965]: _ensure_no_port_binding_failure(port) [ 708.714402] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 708.714402] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 708.715244] env[62965]: nova.exception.PortBindingFailed: Binding failed for port cfb0fc29-145a-45a9-8e88-4fb6176ed1e4, please check neutron logs for more information. [ 708.715244] env[62965]: Removing descriptor: 19 [ 708.715244] env[62965]: ERROR nova.compute.manager [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port cfb0fc29-145a-45a9-8e88-4fb6176ed1e4, please check neutron logs for more information. [ 708.715244] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Traceback (most recent call last): [ 708.715244] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 708.715244] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] yield resources [ 708.715244] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 708.715244] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] self.driver.spawn(context, instance, image_meta, [ 708.715244] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 708.715244] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] self._vmops.spawn(context, instance, image_meta, injected_files, [ 708.715244] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 708.715244] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] vm_ref = self.build_virtual_machine(instance, [ 708.715685] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 708.715685] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] vif_infos = vmwarevif.get_vif_info(self._session, [ 708.715685] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 708.715685] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] for vif in network_info: [ 708.715685] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 708.715685] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] return self._sync_wrapper(fn, *args, **kwargs) [ 708.715685] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 708.715685] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] self.wait() [ 708.715685] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 708.715685] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] self[:] = self._gt.wait() [ 708.715685] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 708.715685] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] return self._exit_event.wait() [ 708.715685] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 708.716094] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] result = hub.switch() [ 708.716094] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 708.716094] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] return self.greenlet.switch() [ 708.716094] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 708.716094] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] result = function(*args, **kwargs) [ 708.716094] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 708.716094] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] return func(*args, **kwargs) [ 708.716094] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 708.716094] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] raise e [ 708.716094] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 708.716094] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] nwinfo = self.network_api.allocate_for_instance( [ 708.716094] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 708.716094] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] created_port_ids = self._update_ports_for_instance( [ 708.716474] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 708.716474] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] with excutils.save_and_reraise_exception(): [ 708.716474] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 708.716474] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] self.force_reraise() [ 708.716474] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 708.716474] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] raise self.value [ 708.716474] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 708.716474] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] updated_port = self._update_port( [ 708.716474] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 708.716474] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] _ensure_no_port_binding_failure(port) [ 708.716474] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 708.716474] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] raise exception.PortBindingFailed(port_id=port['id']) [ 708.716843] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] nova.exception.PortBindingFailed: Binding failed for port cfb0fc29-145a-45a9-8e88-4fb6176ed1e4, please check neutron logs for more information. [ 708.716843] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] [ 708.716843] env[62965]: INFO nova.compute.manager [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Terminating instance [ 708.729141] env[62965]: DEBUG nova.scheduler.client.report [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 708.843858] env[62965]: DEBUG nova.compute.manager [req-4c83c258-fbb4-408a-b264-5da19b24219f req-719724eb-34f5-4408-a2fc-2a1afc26980c service nova] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Received event network-changed-cfb0fc29-145a-45a9-8e88-4fb6176ed1e4 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 708.844124] env[62965]: DEBUG nova.compute.manager [req-4c83c258-fbb4-408a-b264-5da19b24219f req-719724eb-34f5-4408-a2fc-2a1afc26980c service nova] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Refreshing instance network info cache due to event network-changed-cfb0fc29-145a-45a9-8e88-4fb6176ed1e4. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 708.844275] env[62965]: DEBUG oslo_concurrency.lockutils [req-4c83c258-fbb4-408a-b264-5da19b24219f req-719724eb-34f5-4408-a2fc-2a1afc26980c service nova] Acquiring lock "refresh_cache-6c7ed2a3-ed0f-476b-8232-d7a3c27adcee" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.844414] env[62965]: DEBUG oslo_concurrency.lockutils [req-4c83c258-fbb4-408a-b264-5da19b24219f req-719724eb-34f5-4408-a2fc-2a1afc26980c service nova] Acquired lock "refresh_cache-6c7ed2a3-ed0f-476b-8232-d7a3c27adcee" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.844569] env[62965]: DEBUG nova.network.neutron [req-4c83c258-fbb4-408a-b264-5da19b24219f req-719724eb-34f5-4408-a2fc-2a1afc26980c service nova] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Refreshing network info cache for port cfb0fc29-145a-45a9-8e88-4fb6176ed1e4 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 708.866047] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.049997] env[62965]: DEBUG oslo_vmware.api [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389781, 'name': PowerOffVM_Task, 'duration_secs': 0.167356} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.050556] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 709.050788] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 709.051059] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1d8232b8-74f6-4da8-a936-f4fc7bcf21b0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.076737] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 709.077063] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Deleting contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 709.077190] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Deleting the datastore file [datastore1] 02f113d2-f991-4268-b9a4-59e2e6bcf7a6 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 709.077456] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-73ebcee3-5417-4fb2-ab4b-7c2733962163 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.084300] env[62965]: DEBUG oslo_vmware.api [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Waiting for the task: (returnval){ [ 709.084300] env[62965]: value = "task-1389783" [ 709.084300] env[62965]: _type = "Task" [ 709.084300] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.092018] env[62965]: DEBUG oslo_vmware.api [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389783, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.219377] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] Acquiring lock "refresh_cache-6c7ed2a3-ed0f-476b-8232-d7a3c27adcee" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.234661] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.612s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.235310] env[62965]: DEBUG nova.compute.manager [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 709.239020] env[62965]: DEBUG oslo_concurrency.lockutils [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.202s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 709.240074] env[62965]: INFO nova.compute.claims [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 709.363315] env[62965]: DEBUG nova.network.neutron [req-4c83c258-fbb4-408a-b264-5da19b24219f req-719724eb-34f5-4408-a2fc-2a1afc26980c service nova] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 709.415398] env[62965]: DEBUG nova.network.neutron [req-4c83c258-fbb4-408a-b264-5da19b24219f req-719724eb-34f5-4408-a2fc-2a1afc26980c service nova] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.597135] env[62965]: DEBUG oslo_vmware.api [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Task: {'id': task-1389783, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.085299} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.597135] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 709.597135] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Deleted contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 709.597135] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 709.597135] env[62965]: INFO nova.compute.manager [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Took 1.07 seconds to destroy the instance on the hypervisor. [ 709.597462] env[62965]: DEBUG oslo.service.loopingcall [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 709.597462] env[62965]: DEBUG nova.compute.manager [-] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 709.597462] env[62965]: DEBUG nova.network.neutron [-] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 709.611257] env[62965]: DEBUG nova.network.neutron [-] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 709.750813] env[62965]: DEBUG nova.compute.utils [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 709.751602] env[62965]: DEBUG nova.compute.manager [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 709.752020] env[62965]: DEBUG nova.network.neutron [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 709.805419] env[62965]: DEBUG nova.policy [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b4083bef654943e2a8e611c60179772d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c4da028039f34677a49c465a11c57821', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 709.918316] env[62965]: DEBUG oslo_concurrency.lockutils [req-4c83c258-fbb4-408a-b264-5da19b24219f req-719724eb-34f5-4408-a2fc-2a1afc26980c service nova] Releasing lock "refresh_cache-6c7ed2a3-ed0f-476b-8232-d7a3c27adcee" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 709.918721] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] Acquired lock "refresh_cache-6c7ed2a3-ed0f-476b-8232-d7a3c27adcee" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.918912] env[62965]: DEBUG nova.network.neutron [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 710.086753] env[62965]: DEBUG nova.network.neutron [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Successfully created port: 1bd4d220-93d4-492d-9b69-965c342a2559 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 710.113587] env[62965]: DEBUG nova.network.neutron [-] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.257080] env[62965]: DEBUG nova.compute.manager [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 710.444473] env[62965]: DEBUG nova.network.neutron [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 710.518940] env[62965]: DEBUG nova.network.neutron [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.616867] env[62965]: INFO nova.compute.manager [-] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Took 1.02 seconds to deallocate network for instance. [ 710.712172] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd8546f6-c958-4411-bd8f-86ab22d4ad99 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.720538] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb8d8be7-112f-408f-a1a3-d4e5afef8b65 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.752240] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b58f7bd2-f361-4a0d-9677-7eea3cd31bb7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.760323] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4df26b9e-0dab-4152-bb6a-2781813910bf {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.777904] env[62965]: DEBUG nova.compute.provider_tree [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 710.932652] env[62965]: DEBUG nova.compute.manager [req-25bd49e5-44f7-444c-9b97-d99dd6c8f2d9 req-ac55178a-92a8-4d7e-b408-2aa19d90bd64 service nova] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Received event network-vif-deleted-cfb0fc29-145a-45a9-8e88-4fb6176ed1e4 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 711.027205] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] Releasing lock "refresh_cache-6c7ed2a3-ed0f-476b-8232-d7a3c27adcee" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 711.027621] env[62965]: DEBUG nova.compute.manager [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 711.027810] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 711.028584] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1767ed08-a5a0-45eb-bf38-61271ec983a3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.037334] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0380f64-da92-4bfa-8878-811e2bf5aade {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.063795] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee could not be found. [ 711.064237] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 711.064426] env[62965]: INFO nova.compute.manager [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Took 0.04 seconds to destroy the instance on the hypervisor. [ 711.064669] env[62965]: DEBUG oslo.service.loopingcall [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 711.064888] env[62965]: DEBUG nova.compute.manager [-] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 711.064978] env[62965]: DEBUG nova.network.neutron [-] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 711.088660] env[62965]: DEBUG nova.network.neutron [-] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 711.125845] env[62965]: DEBUG oslo_concurrency.lockutils [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 711.202165] env[62965]: ERROR nova.compute.manager [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1bd4d220-93d4-492d-9b69-965c342a2559, please check neutron logs for more information. [ 711.202165] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 711.202165] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 711.202165] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 711.202165] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 711.202165] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 711.202165] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 711.202165] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 711.202165] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 711.202165] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 711.202165] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 711.202165] env[62965]: ERROR nova.compute.manager raise self.value [ 711.202165] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 711.202165] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 711.202165] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 711.202165] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 711.203092] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 711.203092] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 711.203092] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1bd4d220-93d4-492d-9b69-965c342a2559, please check neutron logs for more information. [ 711.203092] env[62965]: ERROR nova.compute.manager [ 711.203092] env[62965]: Traceback (most recent call last): [ 711.203092] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 711.203092] env[62965]: listener.cb(fileno) [ 711.203092] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 711.203092] env[62965]: result = function(*args, **kwargs) [ 711.203092] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 711.203092] env[62965]: return func(*args, **kwargs) [ 711.203092] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 711.203092] env[62965]: raise e [ 711.203092] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 711.203092] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 711.203092] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 711.203092] env[62965]: created_port_ids = self._update_ports_for_instance( [ 711.203092] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 711.203092] env[62965]: with excutils.save_and_reraise_exception(): [ 711.203092] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 711.203092] env[62965]: self.force_reraise() [ 711.203092] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 711.203092] env[62965]: raise self.value [ 711.203092] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 711.203092] env[62965]: updated_port = self._update_port( [ 711.203092] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 711.203092] env[62965]: _ensure_no_port_binding_failure(port) [ 711.203092] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 711.203092] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 711.204648] env[62965]: nova.exception.PortBindingFailed: Binding failed for port 1bd4d220-93d4-492d-9b69-965c342a2559, please check neutron logs for more information. [ 711.204648] env[62965]: Removing descriptor: 19 [ 711.270481] env[62965]: DEBUG nova.compute.manager [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 711.281906] env[62965]: DEBUG nova.scheduler.client.report [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 711.300871] env[62965]: DEBUG nova.virt.hardware [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 711.301156] env[62965]: DEBUG nova.virt.hardware [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 711.301328] env[62965]: DEBUG nova.virt.hardware [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 711.301509] env[62965]: DEBUG nova.virt.hardware [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 711.301654] env[62965]: DEBUG nova.virt.hardware [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 711.301922] env[62965]: DEBUG nova.virt.hardware [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 711.302163] env[62965]: DEBUG nova.virt.hardware [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 711.302325] env[62965]: DEBUG nova.virt.hardware [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 711.302490] env[62965]: DEBUG nova.virt.hardware [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 711.302652] env[62965]: DEBUG nova.virt.hardware [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 711.302826] env[62965]: DEBUG nova.virt.hardware [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 711.303949] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2803aad9-a709-4075-8b13-12ec8687c601 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.312502] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3041ecab-7cf2-4d74-8f12-3d511d076d6d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.326538] env[62965]: ERROR nova.compute.manager [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1bd4d220-93d4-492d-9b69-965c342a2559, please check neutron logs for more information. [ 711.326538] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Traceback (most recent call last): [ 711.326538] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 711.326538] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] yield resources [ 711.326538] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 711.326538] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] self.driver.spawn(context, instance, image_meta, [ 711.326538] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 711.326538] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] self._vmops.spawn(context, instance, image_meta, injected_files, [ 711.326538] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 711.326538] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] vm_ref = self.build_virtual_machine(instance, [ 711.326538] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 711.326931] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] vif_infos = vmwarevif.get_vif_info(self._session, [ 711.326931] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 711.326931] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] for vif in network_info: [ 711.326931] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 711.326931] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] return self._sync_wrapper(fn, *args, **kwargs) [ 711.326931] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 711.326931] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] self.wait() [ 711.326931] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 711.326931] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] self[:] = self._gt.wait() [ 711.326931] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 711.326931] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] return self._exit_event.wait() [ 711.326931] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 711.326931] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] current.throw(*self._exc) [ 711.327319] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 711.327319] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] result = function(*args, **kwargs) [ 711.327319] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 711.327319] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] return func(*args, **kwargs) [ 711.327319] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 711.327319] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] raise e [ 711.327319] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 711.327319] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] nwinfo = self.network_api.allocate_for_instance( [ 711.327319] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 711.327319] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] created_port_ids = self._update_ports_for_instance( [ 711.327319] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 711.327319] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] with excutils.save_and_reraise_exception(): [ 711.327319] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 711.327695] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] self.force_reraise() [ 711.327695] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 711.327695] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] raise self.value [ 711.327695] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 711.327695] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] updated_port = self._update_port( [ 711.327695] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 711.327695] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] _ensure_no_port_binding_failure(port) [ 711.327695] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 711.327695] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] raise exception.PortBindingFailed(port_id=port['id']) [ 711.327695] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] nova.exception.PortBindingFailed: Binding failed for port 1bd4d220-93d4-492d-9b69-965c342a2559, please check neutron logs for more information. [ 711.327695] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] [ 711.327695] env[62965]: INFO nova.compute.manager [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Terminating instance [ 711.591548] env[62965]: DEBUG nova.network.neutron [-] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.789386] env[62965]: DEBUG oslo_concurrency.lockutils [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.550s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.789386] env[62965]: DEBUG nova.compute.manager [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 711.792818] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.588s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 711.794443] env[62965]: INFO nova.compute.claims [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 711.831570] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] Acquiring lock "refresh_cache-852e7cdf-5da3-4217-8960-f3204d4c1035" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.832690] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] Acquired lock "refresh_cache-852e7cdf-5da3-4217-8960-f3204d4c1035" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.832690] env[62965]: DEBUG nova.network.neutron [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 712.095023] env[62965]: INFO nova.compute.manager [-] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Took 1.03 seconds to deallocate network for instance. [ 712.097808] env[62965]: DEBUG nova.compute.claims [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 712.097996] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.304224] env[62965]: DEBUG nova.compute.utils [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 712.305688] env[62965]: DEBUG nova.compute.manager [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 712.306064] env[62965]: DEBUG nova.network.neutron [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 712.359297] env[62965]: DEBUG nova.network.neutron [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.411680] env[62965]: DEBUG nova.policy [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b8fa91537eda4f8a8ce97882c42658f4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e867051e68bb4966a74b4571bff7cc40', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 712.471589] env[62965]: DEBUG nova.network.neutron [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.810549] env[62965]: DEBUG nova.compute.manager [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 712.973850] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] Releasing lock "refresh_cache-852e7cdf-5da3-4217-8960-f3204d4c1035" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.974505] env[62965]: DEBUG nova.compute.manager [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 712.974505] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 712.974809] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-faf29a3a-2dc8-4047-a38f-b6eb93113e17 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.986608] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04e82d25-1df1-49a5-b3c2-3f3ec631a601 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.017029] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 852e7cdf-5da3-4217-8960-f3204d4c1035 could not be found. [ 713.017029] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 713.017150] env[62965]: INFO nova.compute.manager [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Took 0.04 seconds to destroy the instance on the hypervisor. [ 713.017826] env[62965]: DEBUG oslo.service.loopingcall [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 713.017826] env[62965]: DEBUG nova.compute.manager [-] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 713.017826] env[62965]: DEBUG nova.network.neutron [-] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 713.048351] env[62965]: DEBUG nova.network.neutron [-] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 713.075298] env[62965]: DEBUG nova.compute.manager [req-6b535303-d8fc-4eff-8e47-a779241014e1 req-12a86923-953a-4836-bbb5-8a29603f2c91 service nova] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Received event network-changed-1bd4d220-93d4-492d-9b69-965c342a2559 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 713.075298] env[62965]: DEBUG nova.compute.manager [req-6b535303-d8fc-4eff-8e47-a779241014e1 req-12a86923-953a-4836-bbb5-8a29603f2c91 service nova] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Refreshing instance network info cache due to event network-changed-1bd4d220-93d4-492d-9b69-965c342a2559. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 713.075298] env[62965]: DEBUG oslo_concurrency.lockutils [req-6b535303-d8fc-4eff-8e47-a779241014e1 req-12a86923-953a-4836-bbb5-8a29603f2c91 service nova] Acquiring lock "refresh_cache-852e7cdf-5da3-4217-8960-f3204d4c1035" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.075298] env[62965]: DEBUG oslo_concurrency.lockutils [req-6b535303-d8fc-4eff-8e47-a779241014e1 req-12a86923-953a-4836-bbb5-8a29603f2c91 service nova] Acquired lock "refresh_cache-852e7cdf-5da3-4217-8960-f3204d4c1035" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.075298] env[62965]: DEBUG nova.network.neutron [req-6b535303-d8fc-4eff-8e47-a779241014e1 req-12a86923-953a-4836-bbb5-8a29603f2c91 service nova] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Refreshing network info cache for port 1bd4d220-93d4-492d-9b69-965c342a2559 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 713.118492] env[62965]: DEBUG nova.network.neutron [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Successfully created port: 832ab8bd-d991-46d2-b86c-20147968b4fe {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 713.279714] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96918eb5-4909-4b45-ab4c-aa1b682fc930 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.289197] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d44e4b1-a13a-46e7-b828-67f4189a6e16 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.329997] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffd0d5e9-1dcf-43f8-ace2-c20e949ebdbf {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.339978] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-198c5510-4c68-4fc9-a2bf-26c77852a719 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.360265] env[62965]: DEBUG nova.compute.provider_tree [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 713.551871] env[62965]: DEBUG nova.network.neutron [-] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.617911] env[62965]: DEBUG nova.network.neutron [req-6b535303-d8fc-4eff-8e47-a779241014e1 req-12a86923-953a-4836-bbb5-8a29603f2c91 service nova] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 713.706050] env[62965]: DEBUG nova.network.neutron [req-6b535303-d8fc-4eff-8e47-a779241014e1 req-12a86923-953a-4836-bbb5-8a29603f2c91 service nova] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.835062] env[62965]: DEBUG nova.compute.manager [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 713.862085] env[62965]: DEBUG nova.virt.hardware [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 713.862260] env[62965]: DEBUG nova.virt.hardware [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 713.862416] env[62965]: DEBUG nova.virt.hardware [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 713.862591] env[62965]: DEBUG nova.virt.hardware [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 713.862727] env[62965]: DEBUG nova.virt.hardware [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 713.862876] env[62965]: DEBUG nova.virt.hardware [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 713.863088] env[62965]: DEBUG nova.virt.hardware [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 713.863239] env[62965]: DEBUG nova.virt.hardware [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 713.865024] env[62965]: DEBUG nova.virt.hardware [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 713.865024] env[62965]: DEBUG nova.virt.hardware [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 713.865024] env[62965]: DEBUG nova.virt.hardware [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 713.865024] env[62965]: DEBUG nova.scheduler.client.report [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 713.872228] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c179880a-5784-4dba-98f9-4064d2f0e527 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.888137] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90012680-5927-46ce-aecd-5d9dd5ce3645 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.056532] env[62965]: INFO nova.compute.manager [-] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Took 1.04 seconds to deallocate network for instance. [ 714.063141] env[62965]: DEBUG nova.compute.claims [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 714.063328] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 714.209247] env[62965]: DEBUG oslo_concurrency.lockutils [req-6b535303-d8fc-4eff-8e47-a779241014e1 req-12a86923-953a-4836-bbb5-8a29603f2c91 service nova] Releasing lock "refresh_cache-852e7cdf-5da3-4217-8960-f3204d4c1035" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.209941] env[62965]: DEBUG nova.compute.manager [req-6b535303-d8fc-4eff-8e47-a779241014e1 req-12a86923-953a-4836-bbb5-8a29603f2c91 service nova] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Received event network-vif-deleted-1bd4d220-93d4-492d-9b69-965c342a2559 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 714.380424] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.589s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.384252] env[62965]: DEBUG nova.compute.manager [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 714.392588] env[62965]: DEBUG oslo_concurrency.lockutils [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 17.570s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 714.392887] env[62965]: DEBUG nova.objects.instance [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62965) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 714.865599] env[62965]: ERROR nova.compute.manager [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 832ab8bd-d991-46d2-b86c-20147968b4fe, please check neutron logs for more information. [ 714.865599] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 714.865599] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 714.865599] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 714.865599] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 714.865599] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 714.865599] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 714.865599] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 714.865599] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 714.865599] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 714.865599] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 714.865599] env[62965]: ERROR nova.compute.manager raise self.value [ 714.865599] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 714.865599] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 714.865599] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 714.865599] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 714.866482] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 714.866482] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 714.866482] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 832ab8bd-d991-46d2-b86c-20147968b4fe, please check neutron logs for more information. [ 714.866482] env[62965]: ERROR nova.compute.manager [ 714.866482] env[62965]: Traceback (most recent call last): [ 714.866482] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 714.866482] env[62965]: listener.cb(fileno) [ 714.866482] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 714.866482] env[62965]: result = function(*args, **kwargs) [ 714.866482] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 714.866482] env[62965]: return func(*args, **kwargs) [ 714.866482] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 714.866482] env[62965]: raise e [ 714.866482] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 714.866482] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 714.866482] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 714.866482] env[62965]: created_port_ids = self._update_ports_for_instance( [ 714.866482] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 714.866482] env[62965]: with excutils.save_and_reraise_exception(): [ 714.866482] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 714.866482] env[62965]: self.force_reraise() [ 714.866482] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 714.866482] env[62965]: raise self.value [ 714.866482] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 714.866482] env[62965]: updated_port = self._update_port( [ 714.866482] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 714.866482] env[62965]: _ensure_no_port_binding_failure(port) [ 714.866482] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 714.866482] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 714.867563] env[62965]: nova.exception.PortBindingFailed: Binding failed for port 832ab8bd-d991-46d2-b86c-20147968b4fe, please check neutron logs for more information. [ 714.867563] env[62965]: Removing descriptor: 19 [ 714.867563] env[62965]: ERROR nova.compute.manager [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 832ab8bd-d991-46d2-b86c-20147968b4fe, please check neutron logs for more information. [ 714.867563] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Traceback (most recent call last): [ 714.867563] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 714.867563] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] yield resources [ 714.867563] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 714.867563] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] self.driver.spawn(context, instance, image_meta, [ 714.867563] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 714.867563] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] self._vmops.spawn(context, instance, image_meta, injected_files, [ 714.867563] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 714.867563] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] vm_ref = self.build_virtual_machine(instance, [ 714.868197] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 714.868197] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] vif_infos = vmwarevif.get_vif_info(self._session, [ 714.868197] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 714.868197] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] for vif in network_info: [ 714.868197] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 714.868197] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] return self._sync_wrapper(fn, *args, **kwargs) [ 714.868197] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 714.868197] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] self.wait() [ 714.868197] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 714.868197] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] self[:] = self._gt.wait() [ 714.868197] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 714.868197] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] return self._exit_event.wait() [ 714.868197] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 714.868873] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] result = hub.switch() [ 714.868873] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 714.868873] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] return self.greenlet.switch() [ 714.868873] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 714.868873] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] result = function(*args, **kwargs) [ 714.868873] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 714.868873] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] return func(*args, **kwargs) [ 714.868873] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 714.868873] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] raise e [ 714.868873] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 714.868873] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] nwinfo = self.network_api.allocate_for_instance( [ 714.868873] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 714.868873] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] created_port_ids = self._update_ports_for_instance( [ 714.869481] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 714.869481] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] with excutils.save_and_reraise_exception(): [ 714.869481] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 714.869481] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] self.force_reraise() [ 714.869481] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 714.869481] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] raise self.value [ 714.869481] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 714.869481] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] updated_port = self._update_port( [ 714.869481] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 714.869481] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] _ensure_no_port_binding_failure(port) [ 714.869481] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 714.869481] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] raise exception.PortBindingFailed(port_id=port['id']) [ 714.870063] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] nova.exception.PortBindingFailed: Binding failed for port 832ab8bd-d991-46d2-b86c-20147968b4fe, please check neutron logs for more information. [ 714.870063] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] [ 714.870063] env[62965]: INFO nova.compute.manager [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Terminating instance [ 715.065325] env[62965]: DEBUG nova.compute.utils [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 715.069011] env[62965]: DEBUG nova.compute.manager [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 715.069189] env[62965]: DEBUG nova.network.neutron [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 715.119345] env[62965]: DEBUG nova.policy [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '182d38ca18c64bed8f3ecc3f95229756', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1952dedf340f4b7ba0d57af6bba9a749', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 715.347622] env[62965]: DEBUG nova.compute.manager [req-5441c231-5ae2-4326-a5a2-167a22cdf821 req-7881b679-4dba-4e43-a71b-49d51ed6c90c service nova] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Received event network-changed-832ab8bd-d991-46d2-b86c-20147968b4fe {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 715.348462] env[62965]: DEBUG nova.compute.manager [req-5441c231-5ae2-4326-a5a2-167a22cdf821 req-7881b679-4dba-4e43-a71b-49d51ed6c90c service nova] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Refreshing instance network info cache due to event network-changed-832ab8bd-d991-46d2-b86c-20147968b4fe. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 715.348462] env[62965]: DEBUG oslo_concurrency.lockutils [req-5441c231-5ae2-4326-a5a2-167a22cdf821 req-7881b679-4dba-4e43-a71b-49d51ed6c90c service nova] Acquiring lock "refresh_cache-b2f38e92-37a1-4e0b-9292-b20ea43a2038" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.348462] env[62965]: DEBUG oslo_concurrency.lockutils [req-5441c231-5ae2-4326-a5a2-167a22cdf821 req-7881b679-4dba-4e43-a71b-49d51ed6c90c service nova] Acquired lock "refresh_cache-b2f38e92-37a1-4e0b-9292-b20ea43a2038" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.348670] env[62965]: DEBUG nova.network.neutron [req-5441c231-5ae2-4326-a5a2-167a22cdf821 req-7881b679-4dba-4e43-a71b-49d51ed6c90c service nova] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Refreshing network info cache for port 832ab8bd-d991-46d2-b86c-20147968b4fe {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 715.371999] env[62965]: DEBUG oslo_concurrency.lockutils [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] Acquiring lock "refresh_cache-b2f38e92-37a1-4e0b-9292-b20ea43a2038" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.457023] env[62965]: DEBUG nova.network.neutron [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Successfully created port: 84318521-767d-49b3-81f5-02cbb8556c81 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 715.573019] env[62965]: DEBUG nova.compute.manager [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 715.574097] env[62965]: DEBUG oslo_concurrency.lockutils [None req-064d89d1-8b25-486c-9e47-a9bc721a74d6 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.181s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.575249] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.419s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 715.576784] env[62965]: INFO nova.compute.claims [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 715.882287] env[62965]: DEBUG nova.network.neutron [req-5441c231-5ae2-4326-a5a2-167a22cdf821 req-7881b679-4dba-4e43-a71b-49d51ed6c90c service nova] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 715.961909] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] Acquiring lock "474c8faf-cf69-4296-8fba-521e5472c071" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 715.962158] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] Lock "474c8faf-cf69-4296-8fba-521e5472c071" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.018039] env[62965]: DEBUG nova.network.neutron [req-5441c231-5ae2-4326-a5a2-167a22cdf821 req-7881b679-4dba-4e43-a71b-49d51ed6c90c service nova] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.316571] env[62965]: ERROR nova.compute.manager [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 84318521-767d-49b3-81f5-02cbb8556c81, please check neutron logs for more information. [ 716.316571] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 716.316571] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 716.316571] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 716.316571] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 716.316571] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 716.316571] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 716.316571] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 716.316571] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 716.316571] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 716.316571] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 716.316571] env[62965]: ERROR nova.compute.manager raise self.value [ 716.316571] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 716.316571] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 716.316571] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 716.316571] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 716.317064] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 716.317064] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 716.317064] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 84318521-767d-49b3-81f5-02cbb8556c81, please check neutron logs for more information. [ 716.317064] env[62965]: ERROR nova.compute.manager [ 716.317064] env[62965]: Traceback (most recent call last): [ 716.317064] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 716.317064] env[62965]: listener.cb(fileno) [ 716.317064] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 716.317064] env[62965]: result = function(*args, **kwargs) [ 716.317064] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 716.317064] env[62965]: return func(*args, **kwargs) [ 716.317064] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 716.317064] env[62965]: raise e [ 716.317064] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 716.317064] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 716.317064] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 716.317064] env[62965]: created_port_ids = self._update_ports_for_instance( [ 716.317064] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 716.317064] env[62965]: with excutils.save_and_reraise_exception(): [ 716.317064] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 716.317064] env[62965]: self.force_reraise() [ 716.317064] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 716.317064] env[62965]: raise self.value [ 716.317064] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 716.317064] env[62965]: updated_port = self._update_port( [ 716.317064] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 716.317064] env[62965]: _ensure_no_port_binding_failure(port) [ 716.317064] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 716.317064] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 716.317841] env[62965]: nova.exception.PortBindingFailed: Binding failed for port 84318521-767d-49b3-81f5-02cbb8556c81, please check neutron logs for more information. [ 716.317841] env[62965]: Removing descriptor: 19 [ 716.521048] env[62965]: DEBUG oslo_concurrency.lockutils [req-5441c231-5ae2-4326-a5a2-167a22cdf821 req-7881b679-4dba-4e43-a71b-49d51ed6c90c service nova] Releasing lock "refresh_cache-b2f38e92-37a1-4e0b-9292-b20ea43a2038" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.521355] env[62965]: DEBUG nova.compute.manager [req-5441c231-5ae2-4326-a5a2-167a22cdf821 req-7881b679-4dba-4e43-a71b-49d51ed6c90c service nova] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Received event network-vif-deleted-832ab8bd-d991-46d2-b86c-20147968b4fe {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 716.521709] env[62965]: DEBUG oslo_concurrency.lockutils [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] Acquired lock "refresh_cache-b2f38e92-37a1-4e0b-9292-b20ea43a2038" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.521880] env[62965]: DEBUG nova.network.neutron [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 716.585828] env[62965]: DEBUG nova.compute.manager [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 716.613136] env[62965]: DEBUG nova.virt.hardware [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 716.613801] env[62965]: DEBUG nova.virt.hardware [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 716.613801] env[62965]: DEBUG nova.virt.hardware [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 716.613801] env[62965]: DEBUG nova.virt.hardware [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 716.613975] env[62965]: DEBUG nova.virt.hardware [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 716.613975] env[62965]: DEBUG nova.virt.hardware [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 716.614175] env[62965]: DEBUG nova.virt.hardware [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 716.614332] env[62965]: DEBUG nova.virt.hardware [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 716.614494] env[62965]: DEBUG nova.virt.hardware [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 716.614652] env[62965]: DEBUG nova.virt.hardware [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 716.614820] env[62965]: DEBUG nova.virt.hardware [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 716.616195] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-937a6031-a97c-4a30-ba20-6cfbf3ada9c5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.624139] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee895b85-2658-459c-8086-28f0fa8aa80e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.640038] env[62965]: ERROR nova.compute.manager [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 84318521-767d-49b3-81f5-02cbb8556c81, please check neutron logs for more information. [ 716.640038] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Traceback (most recent call last): [ 716.640038] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 716.640038] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] yield resources [ 716.640038] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 716.640038] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] self.driver.spawn(context, instance, image_meta, [ 716.640038] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 716.640038] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] self._vmops.spawn(context, instance, image_meta, injected_files, [ 716.640038] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 716.640038] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] vm_ref = self.build_virtual_machine(instance, [ 716.640038] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 716.640371] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] vif_infos = vmwarevif.get_vif_info(self._session, [ 716.640371] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 716.640371] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] for vif in network_info: [ 716.640371] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 716.640371] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] return self._sync_wrapper(fn, *args, **kwargs) [ 716.640371] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 716.640371] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] self.wait() [ 716.640371] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 716.640371] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] self[:] = self._gt.wait() [ 716.640371] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 716.640371] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] return self._exit_event.wait() [ 716.640371] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 716.640371] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] current.throw(*self._exc) [ 716.640762] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 716.640762] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] result = function(*args, **kwargs) [ 716.640762] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 716.640762] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] return func(*args, **kwargs) [ 716.640762] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 716.640762] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] raise e [ 716.640762] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 716.640762] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] nwinfo = self.network_api.allocate_for_instance( [ 716.640762] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 716.640762] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] created_port_ids = self._update_ports_for_instance( [ 716.640762] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 716.640762] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] with excutils.save_and_reraise_exception(): [ 716.640762] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 716.641166] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] self.force_reraise() [ 716.641166] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 716.641166] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] raise self.value [ 716.641166] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 716.641166] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] updated_port = self._update_port( [ 716.641166] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 716.641166] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] _ensure_no_port_binding_failure(port) [ 716.641166] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 716.641166] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] raise exception.PortBindingFailed(port_id=port['id']) [ 716.641166] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] nova.exception.PortBindingFailed: Binding failed for port 84318521-767d-49b3-81f5-02cbb8556c81, please check neutron logs for more information. [ 716.641166] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] [ 716.641166] env[62965]: INFO nova.compute.manager [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Terminating instance [ 716.911166] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caf55102-b569-4e93-b19d-a8de67f30cf7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.918714] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed895b78-0eff-4d9e-9d6f-fe453413f94f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.951040] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b56b0547-20cd-468d-b01b-66605701459f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.958780] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52347130-f4f0-4867-8c00-b1f80a0bfa42 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.972429] env[62965]: DEBUG nova.compute.provider_tree [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 717.040328] env[62965]: DEBUG nova.network.neutron [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 717.118390] env[62965]: DEBUG nova.network.neutron [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.146875] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "refresh_cache-83c4b731-780a-4f3c-bbdd-b624a52365ef" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.146875] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquired lock "refresh_cache-83c4b731-780a-4f3c-bbdd-b624a52365ef" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.146875] env[62965]: DEBUG nova.network.neutron [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 717.379081] env[62965]: DEBUG nova.compute.manager [req-a7f4dcd1-e5f3-47cb-a169-9b718a071d66 req-18d5cbd0-ae14-4c6f-beab-f8d61874639c service nova] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Received event network-changed-84318521-767d-49b3-81f5-02cbb8556c81 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 717.379236] env[62965]: DEBUG nova.compute.manager [req-a7f4dcd1-e5f3-47cb-a169-9b718a071d66 req-18d5cbd0-ae14-4c6f-beab-f8d61874639c service nova] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Refreshing instance network info cache due to event network-changed-84318521-767d-49b3-81f5-02cbb8556c81. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 717.379421] env[62965]: DEBUG oslo_concurrency.lockutils [req-a7f4dcd1-e5f3-47cb-a169-9b718a071d66 req-18d5cbd0-ae14-4c6f-beab-f8d61874639c service nova] Acquiring lock "refresh_cache-83c4b731-780a-4f3c-bbdd-b624a52365ef" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.476927] env[62965]: DEBUG nova.scheduler.client.report [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 717.621485] env[62965]: DEBUG oslo_concurrency.lockutils [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] Releasing lock "refresh_cache-b2f38e92-37a1-4e0b-9292-b20ea43a2038" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.623283] env[62965]: DEBUG nova.compute.manager [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 717.623283] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 717.623283] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8d3d97f1-4e8f-4183-87db-91c4d6a5f9fc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.631344] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f57ecdcc-d663-49ff-898d-cd2f0ab01613 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.655167] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b2f38e92-37a1-4e0b-9292-b20ea43a2038 could not be found. [ 717.655389] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 717.655567] env[62965]: INFO nova.compute.manager [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Took 0.03 seconds to destroy the instance on the hypervisor. [ 717.655845] env[62965]: DEBUG oslo.service.loopingcall [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 717.656094] env[62965]: DEBUG nova.compute.manager [-] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 717.656126] env[62965]: DEBUG nova.network.neutron [-] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 717.672029] env[62965]: DEBUG nova.network.neutron [-] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 717.673808] env[62965]: DEBUG nova.network.neutron [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 717.743621] env[62965]: DEBUG nova.network.neutron [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.982169] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.407s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.982703] env[62965]: DEBUG nova.compute.manager [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 717.985742] env[62965]: DEBUG oslo_concurrency.lockutils [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.813s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.987201] env[62965]: INFO nova.compute.claims [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 718.179029] env[62965]: DEBUG nova.network.neutron [-] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.246679] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Releasing lock "refresh_cache-83c4b731-780a-4f3c-bbdd-b624a52365ef" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.247122] env[62965]: DEBUG nova.compute.manager [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 718.247318] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 718.248036] env[62965]: DEBUG oslo_concurrency.lockutils [req-a7f4dcd1-e5f3-47cb-a169-9b718a071d66 req-18d5cbd0-ae14-4c6f-beab-f8d61874639c service nova] Acquired lock "refresh_cache-83c4b731-780a-4f3c-bbdd-b624a52365ef" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.248036] env[62965]: DEBUG nova.network.neutron [req-a7f4dcd1-e5f3-47cb-a169-9b718a071d66 req-18d5cbd0-ae14-4c6f-beab-f8d61874639c service nova] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Refreshing network info cache for port 84318521-767d-49b3-81f5-02cbb8556c81 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 718.248911] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c3404586-b60c-426a-a00b-8d53f2051548 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.259476] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bee46c88-6ebd-4803-a2c2-a9f14c81b23d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.281298] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 83c4b731-780a-4f3c-bbdd-b624a52365ef could not be found. [ 718.281518] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 718.281694] env[62965]: INFO nova.compute.manager [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Took 0.03 seconds to destroy the instance on the hypervisor. [ 718.281934] env[62965]: DEBUG oslo.service.loopingcall [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 718.282173] env[62965]: DEBUG nova.compute.manager [-] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 718.282267] env[62965]: DEBUG nova.network.neutron [-] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 718.297771] env[62965]: DEBUG nova.network.neutron [-] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 718.487876] env[62965]: DEBUG nova.compute.utils [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 718.489305] env[62965]: DEBUG nova.compute.manager [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 718.489453] env[62965]: DEBUG nova.network.neutron [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 718.536420] env[62965]: DEBUG nova.policy [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3d68ea75f03f4d2f99cc81efac3c0a8d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e40c42af10c1482783c6f21c0d0c8227', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 718.682212] env[62965]: INFO nova.compute.manager [-] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Took 1.03 seconds to deallocate network for instance. [ 718.684547] env[62965]: DEBUG nova.compute.claims [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 718.684720] env[62965]: DEBUG oslo_concurrency.lockutils [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.766693] env[62965]: DEBUG nova.network.neutron [req-a7f4dcd1-e5f3-47cb-a169-9b718a071d66 req-18d5cbd0-ae14-4c6f-beab-f8d61874639c service nova] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 718.779688] env[62965]: DEBUG nova.network.neutron [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Successfully created port: 57373793-dcc9-4a89-8e78-d8a737cacada {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 718.799861] env[62965]: DEBUG nova.network.neutron [-] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.842755] env[62965]: DEBUG nova.network.neutron [req-a7f4dcd1-e5f3-47cb-a169-9b718a071d66 req-18d5cbd0-ae14-4c6f-beab-f8d61874639c service nova] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.992139] env[62965]: DEBUG nova.compute.manager [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 719.302496] env[62965]: INFO nova.compute.manager [-] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Took 1.02 seconds to deallocate network for instance. [ 719.305558] env[62965]: DEBUG nova.compute.claims [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 719.305900] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.346693] env[62965]: DEBUG oslo_concurrency.lockutils [req-a7f4dcd1-e5f3-47cb-a169-9b718a071d66 req-18d5cbd0-ae14-4c6f-beab-f8d61874639c service nova] Releasing lock "refresh_cache-83c4b731-780a-4f3c-bbdd-b624a52365ef" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.347022] env[62965]: DEBUG nova.compute.manager [req-a7f4dcd1-e5f3-47cb-a169-9b718a071d66 req-18d5cbd0-ae14-4c6f-beab-f8d61874639c service nova] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Received event network-vif-deleted-84318521-767d-49b3-81f5-02cbb8556c81 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 719.409082] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c661038c-6051-47fa-bdc8-1714d6d528a9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.416819] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d5f2cbb-1848-4dbe-8bb9-72481e530973 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.450512] env[62965]: DEBUG nova.compute.manager [req-8c82c71e-f2e5-40a9-82be-151095359a3a req-e8bb1adb-3c8f-47c3-9aa0-d3ade652b0c4 service nova] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Received event network-changed-57373793-dcc9-4a89-8e78-d8a737cacada {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 719.450700] env[62965]: DEBUG nova.compute.manager [req-8c82c71e-f2e5-40a9-82be-151095359a3a req-e8bb1adb-3c8f-47c3-9aa0-d3ade652b0c4 service nova] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Refreshing instance network info cache due to event network-changed-57373793-dcc9-4a89-8e78-d8a737cacada. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 719.450925] env[62965]: DEBUG oslo_concurrency.lockutils [req-8c82c71e-f2e5-40a9-82be-151095359a3a req-e8bb1adb-3c8f-47c3-9aa0-d3ade652b0c4 service nova] Acquiring lock "refresh_cache-f8ba32e4-6234-4dee-8bc5-b7a12105a3c4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.451088] env[62965]: DEBUG oslo_concurrency.lockutils [req-8c82c71e-f2e5-40a9-82be-151095359a3a req-e8bb1adb-3c8f-47c3-9aa0-d3ade652b0c4 service nova] Acquired lock "refresh_cache-f8ba32e4-6234-4dee-8bc5-b7a12105a3c4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.451250] env[62965]: DEBUG nova.network.neutron [req-8c82c71e-f2e5-40a9-82be-151095359a3a req-e8bb1adb-3c8f-47c3-9aa0-d3ade652b0c4 service nova] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Refreshing network info cache for port 57373793-dcc9-4a89-8e78-d8a737cacada {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 719.452643] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79788367-24e7-490d-a277-17a658995a2e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.462860] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d848863-61e4-48c0-baf7-e377eb8ff135 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.476333] env[62965]: DEBUG nova.compute.provider_tree [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 719.669923] env[62965]: ERROR nova.compute.manager [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 57373793-dcc9-4a89-8e78-d8a737cacada, please check neutron logs for more information. [ 719.669923] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 719.669923] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 719.669923] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 719.669923] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 719.669923] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 719.669923] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 719.669923] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 719.669923] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 719.669923] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 719.669923] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 719.669923] env[62965]: ERROR nova.compute.manager raise self.value [ 719.669923] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 719.669923] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 719.669923] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 719.669923] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 719.670474] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 719.670474] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 719.670474] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 57373793-dcc9-4a89-8e78-d8a737cacada, please check neutron logs for more information. [ 719.670474] env[62965]: ERROR nova.compute.manager [ 719.670474] env[62965]: Traceback (most recent call last): [ 719.670474] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 719.670474] env[62965]: listener.cb(fileno) [ 719.670474] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 719.670474] env[62965]: result = function(*args, **kwargs) [ 719.670474] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 719.670474] env[62965]: return func(*args, **kwargs) [ 719.670474] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 719.670474] env[62965]: raise e [ 719.670474] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 719.670474] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 719.670474] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 719.670474] env[62965]: created_port_ids = self._update_ports_for_instance( [ 719.670474] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 719.670474] env[62965]: with excutils.save_and_reraise_exception(): [ 719.670474] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 719.670474] env[62965]: self.force_reraise() [ 719.670474] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 719.670474] env[62965]: raise self.value [ 719.670474] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 719.670474] env[62965]: updated_port = self._update_port( [ 719.670474] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 719.670474] env[62965]: _ensure_no_port_binding_failure(port) [ 719.670474] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 719.670474] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 719.671353] env[62965]: nova.exception.PortBindingFailed: Binding failed for port 57373793-dcc9-4a89-8e78-d8a737cacada, please check neutron logs for more information. [ 719.671353] env[62965]: Removing descriptor: 19 [ 719.973985] env[62965]: DEBUG nova.network.neutron [req-8c82c71e-f2e5-40a9-82be-151095359a3a req-e8bb1adb-3c8f-47c3-9aa0-d3ade652b0c4 service nova] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.979086] env[62965]: DEBUG nova.scheduler.client.report [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 720.002482] env[62965]: DEBUG nova.compute.manager [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 720.037294] env[62965]: DEBUG nova.virt.hardware [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 720.037543] env[62965]: DEBUG nova.virt.hardware [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 720.037786] env[62965]: DEBUG nova.virt.hardware [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 720.037987] env[62965]: DEBUG nova.virt.hardware [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 720.038205] env[62965]: DEBUG nova.virt.hardware [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 720.040098] env[62965]: DEBUG nova.virt.hardware [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 720.040339] env[62965]: DEBUG nova.virt.hardware [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 720.040506] env[62965]: DEBUG nova.virt.hardware [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 720.040710] env[62965]: DEBUG nova.virt.hardware [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 720.041013] env[62965]: DEBUG nova.virt.hardware [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 720.041691] env[62965]: DEBUG nova.virt.hardware [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 720.043433] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac58ca27-6be0-40b0-b0a2-c93b572d52de {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.054248] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d19965e5-f2bb-4816-9922-785e1b92f2e3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.063518] env[62965]: DEBUG nova.network.neutron [req-8c82c71e-f2e5-40a9-82be-151095359a3a req-e8bb1adb-3c8f-47c3-9aa0-d3ade652b0c4 service nova] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.079986] env[62965]: ERROR nova.compute.manager [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 57373793-dcc9-4a89-8e78-d8a737cacada, please check neutron logs for more information. [ 720.079986] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Traceback (most recent call last): [ 720.079986] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 720.079986] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] yield resources [ 720.079986] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 720.079986] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] self.driver.spawn(context, instance, image_meta, [ 720.079986] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 720.079986] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 720.079986] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 720.079986] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] vm_ref = self.build_virtual_machine(instance, [ 720.079986] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 720.080696] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] vif_infos = vmwarevif.get_vif_info(self._session, [ 720.080696] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 720.080696] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] for vif in network_info: [ 720.080696] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 720.080696] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] return self._sync_wrapper(fn, *args, **kwargs) [ 720.080696] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 720.080696] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] self.wait() [ 720.080696] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 720.080696] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] self[:] = self._gt.wait() [ 720.080696] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 720.080696] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] return self._exit_event.wait() [ 720.080696] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 720.080696] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] current.throw(*self._exc) [ 720.081448] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 720.081448] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] result = function(*args, **kwargs) [ 720.081448] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 720.081448] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] return func(*args, **kwargs) [ 720.081448] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 720.081448] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] raise e [ 720.081448] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 720.081448] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] nwinfo = self.network_api.allocate_for_instance( [ 720.081448] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 720.081448] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] created_port_ids = self._update_ports_for_instance( [ 720.081448] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 720.081448] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] with excutils.save_and_reraise_exception(): [ 720.081448] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.081966] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] self.force_reraise() [ 720.081966] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.081966] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] raise self.value [ 720.081966] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 720.081966] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] updated_port = self._update_port( [ 720.081966] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.081966] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] _ensure_no_port_binding_failure(port) [ 720.081966] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.081966] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] raise exception.PortBindingFailed(port_id=port['id']) [ 720.081966] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] nova.exception.PortBindingFailed: Binding failed for port 57373793-dcc9-4a89-8e78-d8a737cacada, please check neutron logs for more information. [ 720.081966] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] [ 720.081966] env[62965]: INFO nova.compute.manager [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Terminating instance [ 720.484341] env[62965]: DEBUG oslo_concurrency.lockutils [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.498s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.484864] env[62965]: DEBUG nova.compute.manager [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 720.487635] env[62965]: DEBUG oslo_concurrency.lockutils [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.288s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.489142] env[62965]: INFO nova.compute.claims [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 720.565572] env[62965]: DEBUG oslo_concurrency.lockutils [req-8c82c71e-f2e5-40a9-82be-151095359a3a req-e8bb1adb-3c8f-47c3-9aa0-d3ade652b0c4 service nova] Releasing lock "refresh_cache-f8ba32e4-6234-4dee-8bc5-b7a12105a3c4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.586883] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] Acquiring lock "refresh_cache-f8ba32e4-6234-4dee-8bc5-b7a12105a3c4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.587080] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] Acquired lock "refresh_cache-f8ba32e4-6234-4dee-8bc5-b7a12105a3c4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.587262] env[62965]: DEBUG nova.network.neutron [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 720.993325] env[62965]: DEBUG nova.compute.utils [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 720.996571] env[62965]: DEBUG nova.compute.manager [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 720.996668] env[62965]: DEBUG nova.network.neutron [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 721.045767] env[62965]: DEBUG nova.policy [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7e9d1908a194480e9ee9f5fa32b15cc6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd9897f889b854681909285c643d1ae86', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 721.104041] env[62965]: DEBUG nova.network.neutron [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.212750] env[62965]: DEBUG nova.network.neutron [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.342319] env[62965]: DEBUG nova.network.neutron [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Successfully created port: 46f93633-3617-45b5-9054-fa6bfc49bee8 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 721.502714] env[62965]: DEBUG nova.compute.manager [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 721.509673] env[62965]: DEBUG nova.compute.manager [req-bf302c9c-370e-441b-8426-c0367db9a837 req-37935dc9-61a8-4fd9-bbf5-d5001c1c7766 service nova] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Received event network-vif-deleted-57373793-dcc9-4a89-8e78-d8a737cacada {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 721.715809] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] Releasing lock "refresh_cache-f8ba32e4-6234-4dee-8bc5-b7a12105a3c4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.716314] env[62965]: DEBUG nova.compute.manager [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 721.717942] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 721.717942] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-51456911-d302-45dd-9909-900072efc8a3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.726235] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9512992f-4d32-4c35-a673-d4fab4156dcb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.761603] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f8ba32e4-6234-4dee-8bc5-b7a12105a3c4 could not be found. [ 721.761940] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 721.762273] env[62965]: INFO nova.compute.manager [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Took 0.05 seconds to destroy the instance on the hypervisor. [ 721.762664] env[62965]: DEBUG oslo.service.loopingcall [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 721.766759] env[62965]: DEBUG nova.compute.manager [-] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 721.766927] env[62965]: DEBUG nova.network.neutron [-] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 721.791901] env[62965]: DEBUG nova.network.neutron [-] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.932882] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8cd4e6b-8c86-4771-8bc5-f1cd1a53cde1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.940915] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fce99328-51bd-45fc-9f21-7d670ff9c2a5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.971595] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5046fd7e-3ebc-4e49-a2c7-1d688d8d2a55 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.978564] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8ea1073-4c8a-408b-96ab-c451f2fc5060 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.991622] env[62965]: DEBUG nova.compute.provider_tree [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 722.294119] env[62965]: DEBUG nova.network.neutron [-] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.366844] env[62965]: ERROR nova.compute.manager [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 46f93633-3617-45b5-9054-fa6bfc49bee8, please check neutron logs for more information. [ 722.366844] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 722.366844] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 722.366844] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 722.366844] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 722.366844] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 722.366844] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 722.366844] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 722.366844] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 722.366844] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 722.366844] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 722.366844] env[62965]: ERROR nova.compute.manager raise self.value [ 722.366844] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 722.366844] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 722.366844] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 722.366844] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 722.367400] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 722.367400] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 722.367400] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 46f93633-3617-45b5-9054-fa6bfc49bee8, please check neutron logs for more information. [ 722.367400] env[62965]: ERROR nova.compute.manager [ 722.367400] env[62965]: Traceback (most recent call last): [ 722.367400] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 722.367400] env[62965]: listener.cb(fileno) [ 722.367400] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 722.367400] env[62965]: result = function(*args, **kwargs) [ 722.367400] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 722.367400] env[62965]: return func(*args, **kwargs) [ 722.367400] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 722.367400] env[62965]: raise e [ 722.367400] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 722.367400] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 722.367400] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 722.367400] env[62965]: created_port_ids = self._update_ports_for_instance( [ 722.367400] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 722.367400] env[62965]: with excutils.save_and_reraise_exception(): [ 722.367400] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 722.367400] env[62965]: self.force_reraise() [ 722.367400] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 722.367400] env[62965]: raise self.value [ 722.367400] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 722.367400] env[62965]: updated_port = self._update_port( [ 722.367400] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 722.367400] env[62965]: _ensure_no_port_binding_failure(port) [ 722.367400] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 722.367400] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 722.368305] env[62965]: nova.exception.PortBindingFailed: Binding failed for port 46f93633-3617-45b5-9054-fa6bfc49bee8, please check neutron logs for more information. [ 722.368305] env[62965]: Removing descriptor: 19 [ 722.494720] env[62965]: DEBUG nova.scheduler.client.report [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 722.514248] env[62965]: DEBUG nova.compute.manager [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 722.538765] env[62965]: DEBUG nova.virt.hardware [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 722.539028] env[62965]: DEBUG nova.virt.hardware [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 722.539190] env[62965]: DEBUG nova.virt.hardware [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 722.539373] env[62965]: DEBUG nova.virt.hardware [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 722.539518] env[62965]: DEBUG nova.virt.hardware [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 722.539687] env[62965]: DEBUG nova.virt.hardware [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 722.539897] env[62965]: DEBUG nova.virt.hardware [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 722.540067] env[62965]: DEBUG nova.virt.hardware [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 722.540237] env[62965]: DEBUG nova.virt.hardware [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 722.540395] env[62965]: DEBUG nova.virt.hardware [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 722.540563] env[62965]: DEBUG nova.virt.hardware [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 722.541664] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73ca9731-057e-4284-b76c-be8142de87db {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.550246] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-452d7f93-2b84-4095-874d-057053aef0a3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.563621] env[62965]: ERROR nova.compute.manager [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 46f93633-3617-45b5-9054-fa6bfc49bee8, please check neutron logs for more information. [ 722.563621] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Traceback (most recent call last): [ 722.563621] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 722.563621] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] yield resources [ 722.563621] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 722.563621] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] self.driver.spawn(context, instance, image_meta, [ 722.563621] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 722.563621] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 722.563621] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 722.563621] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] vm_ref = self.build_virtual_machine(instance, [ 722.563621] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 722.564032] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] vif_infos = vmwarevif.get_vif_info(self._session, [ 722.564032] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 722.564032] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] for vif in network_info: [ 722.564032] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 722.564032] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] return self._sync_wrapper(fn, *args, **kwargs) [ 722.564032] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 722.564032] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] self.wait() [ 722.564032] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 722.564032] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] self[:] = self._gt.wait() [ 722.564032] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 722.564032] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] return self._exit_event.wait() [ 722.564032] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 722.564032] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] current.throw(*self._exc) [ 722.564448] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 722.564448] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] result = function(*args, **kwargs) [ 722.564448] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 722.564448] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] return func(*args, **kwargs) [ 722.564448] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 722.564448] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] raise e [ 722.564448] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 722.564448] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] nwinfo = self.network_api.allocate_for_instance( [ 722.564448] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 722.564448] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] created_port_ids = self._update_ports_for_instance( [ 722.564448] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 722.564448] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] with excutils.save_and_reraise_exception(): [ 722.564448] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 722.564866] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] self.force_reraise() [ 722.564866] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 722.564866] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] raise self.value [ 722.564866] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 722.564866] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] updated_port = self._update_port( [ 722.564866] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 722.564866] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] _ensure_no_port_binding_failure(port) [ 722.564866] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 722.564866] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] raise exception.PortBindingFailed(port_id=port['id']) [ 722.564866] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] nova.exception.PortBindingFailed: Binding failed for port 46f93633-3617-45b5-9054-fa6bfc49bee8, please check neutron logs for more information. [ 722.564866] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] [ 722.564866] env[62965]: INFO nova.compute.manager [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Terminating instance [ 722.799416] env[62965]: INFO nova.compute.manager [-] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Took 1.03 seconds to deallocate network for instance. [ 722.801896] env[62965]: DEBUG nova.compute.claims [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 722.802096] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.999625] env[62965]: DEBUG oslo_concurrency.lockutils [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.512s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.000174] env[62965]: DEBUG nova.compute.manager [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 723.002774] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.761s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.004754] env[62965]: INFO nova.compute.claims [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 723.067680] env[62965]: DEBUG oslo_concurrency.lockutils [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Acquiring lock "refresh_cache-ece118fc-a39e-443a-bfd1-fc332da2452f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.067954] env[62965]: DEBUG oslo_concurrency.lockutils [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Acquired lock "refresh_cache-ece118fc-a39e-443a-bfd1-fc332da2452f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.068169] env[62965]: DEBUG nova.network.neutron [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 723.511614] env[62965]: DEBUG nova.compute.utils [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 723.512741] env[62965]: DEBUG nova.compute.manager [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 723.512898] env[62965]: DEBUG nova.network.neutron [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 723.549757] env[62965]: DEBUG nova.compute.manager [req-b2996e62-d5f5-4ceb-83be-a998f90b3f49 req-2cdca229-4ca2-4a1b-ad86-59696fb52218 service nova] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Received event network-changed-46f93633-3617-45b5-9054-fa6bfc49bee8 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 723.549967] env[62965]: DEBUG nova.compute.manager [req-b2996e62-d5f5-4ceb-83be-a998f90b3f49 req-2cdca229-4ca2-4a1b-ad86-59696fb52218 service nova] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Refreshing instance network info cache due to event network-changed-46f93633-3617-45b5-9054-fa6bfc49bee8. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 723.551152] env[62965]: DEBUG oslo_concurrency.lockutils [req-b2996e62-d5f5-4ceb-83be-a998f90b3f49 req-2cdca229-4ca2-4a1b-ad86-59696fb52218 service nova] Acquiring lock "refresh_cache-ece118fc-a39e-443a-bfd1-fc332da2452f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.566384] env[62965]: DEBUG nova.policy [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4c6cdbd0473c4feaa68cb973f6da7e37', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '502d87d3c6b74442a37e2055c7bca243', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 723.596309] env[62965]: DEBUG nova.network.neutron [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 723.702231] env[62965]: DEBUG nova.network.neutron [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.835848] env[62965]: DEBUG nova.network.neutron [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Successfully created port: 046d1077-d6d9-4ae5-9a33-89a4ba0a7b08 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 724.016608] env[62965]: DEBUG nova.compute.manager [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 724.204868] env[62965]: DEBUG oslo_concurrency.lockutils [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Releasing lock "refresh_cache-ece118fc-a39e-443a-bfd1-fc332da2452f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.205297] env[62965]: DEBUG nova.compute.manager [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 724.205481] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 724.205783] env[62965]: DEBUG oslo_concurrency.lockutils [req-b2996e62-d5f5-4ceb-83be-a998f90b3f49 req-2cdca229-4ca2-4a1b-ad86-59696fb52218 service nova] Acquired lock "refresh_cache-ece118fc-a39e-443a-bfd1-fc332da2452f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.205949] env[62965]: DEBUG nova.network.neutron [req-b2996e62-d5f5-4ceb-83be-a998f90b3f49 req-2cdca229-4ca2-4a1b-ad86-59696fb52218 service nova] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Refreshing network info cache for port 46f93633-3617-45b5-9054-fa6bfc49bee8 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 724.206965] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-94115130-81a7-4ac9-8b9f-e3f0c8b692ae {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.216419] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8a1ff12-dd57-4768-a7f5-ca70503c891b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.243599] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ece118fc-a39e-443a-bfd1-fc332da2452f could not be found. [ 724.243866] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 724.244079] env[62965]: INFO nova.compute.manager [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 724.244349] env[62965]: DEBUG oslo.service.loopingcall [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 724.247010] env[62965]: DEBUG nova.compute.manager [-] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 724.247139] env[62965]: DEBUG nova.network.neutron [-] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 724.265195] env[62965]: DEBUG nova.network.neutron [-] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 724.401818] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a78387ef-525f-4c0c-9565-2ae6d2ec2a4d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.409792] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9726018a-4f45-4e28-83a0-7dfd8a61cb52 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.440461] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea21c07a-6f5f-43f9-b14f-3ab922680126 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.448541] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc6c892e-e4a1-41c7-8c49-7b23a0fc19d0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.462021] env[62965]: DEBUG nova.compute.provider_tree [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 724.713578] env[62965]: ERROR nova.compute.manager [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 046d1077-d6d9-4ae5-9a33-89a4ba0a7b08, please check neutron logs for more information. [ 724.713578] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 724.713578] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 724.713578] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 724.713578] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 724.713578] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 724.713578] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 724.713578] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 724.713578] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.713578] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 724.713578] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.713578] env[62965]: ERROR nova.compute.manager raise self.value [ 724.713578] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 724.713578] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 724.713578] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.713578] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 724.714270] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.714270] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 724.714270] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 046d1077-d6d9-4ae5-9a33-89a4ba0a7b08, please check neutron logs for more information. [ 724.714270] env[62965]: ERROR nova.compute.manager [ 724.714270] env[62965]: Traceback (most recent call last): [ 724.714270] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 724.714270] env[62965]: listener.cb(fileno) [ 724.714270] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 724.714270] env[62965]: result = function(*args, **kwargs) [ 724.714270] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 724.714270] env[62965]: return func(*args, **kwargs) [ 724.714270] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 724.714270] env[62965]: raise e [ 724.714270] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 724.714270] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 724.714270] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 724.714270] env[62965]: created_port_ids = self._update_ports_for_instance( [ 724.714270] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 724.714270] env[62965]: with excutils.save_and_reraise_exception(): [ 724.714270] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 724.714270] env[62965]: self.force_reraise() [ 724.714270] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 724.714270] env[62965]: raise self.value [ 724.714270] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 724.714270] env[62965]: updated_port = self._update_port( [ 724.714270] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 724.714270] env[62965]: _ensure_no_port_binding_failure(port) [ 724.714270] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 724.714270] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 724.715077] env[62965]: nova.exception.PortBindingFailed: Binding failed for port 046d1077-d6d9-4ae5-9a33-89a4ba0a7b08, please check neutron logs for more information. [ 724.715077] env[62965]: Removing descriptor: 19 [ 724.768121] env[62965]: DEBUG nova.network.neutron [-] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.784274] env[62965]: DEBUG nova.network.neutron [req-b2996e62-d5f5-4ceb-83be-a998f90b3f49 req-2cdca229-4ca2-4a1b-ad86-59696fb52218 service nova] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 724.906367] env[62965]: DEBUG nova.network.neutron [req-b2996e62-d5f5-4ceb-83be-a998f90b3f49 req-2cdca229-4ca2-4a1b-ad86-59696fb52218 service nova] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.965476] env[62965]: DEBUG nova.scheduler.client.report [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 725.028890] env[62965]: DEBUG nova.compute.manager [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 725.057791] env[62965]: DEBUG nova.virt.hardware [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 725.058356] env[62965]: DEBUG nova.virt.hardware [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 725.058811] env[62965]: DEBUG nova.virt.hardware [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 725.059281] env[62965]: DEBUG nova.virt.hardware [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 725.059673] env[62965]: DEBUG nova.virt.hardware [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 725.059992] env[62965]: DEBUG nova.virt.hardware [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 725.060332] env[62965]: DEBUG nova.virt.hardware [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 725.060612] env[62965]: DEBUG nova.virt.hardware [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 725.060914] env[62965]: DEBUG nova.virt.hardware [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 725.061220] env[62965]: DEBUG nova.virt.hardware [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 725.061507] env[62965]: DEBUG nova.virt.hardware [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 725.062746] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f4a0b7f-8bc5-45d1-892c-9ad1c08ccd8f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.075018] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f76eced-330c-4ae7-a8dc-1b780445108f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.085020] env[62965]: ERROR nova.compute.manager [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 046d1077-d6d9-4ae5-9a33-89a4ba0a7b08, please check neutron logs for more information. [ 725.085020] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Traceback (most recent call last): [ 725.085020] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 725.085020] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] yield resources [ 725.085020] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 725.085020] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] self.driver.spawn(context, instance, image_meta, [ 725.085020] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 725.085020] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 725.085020] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 725.085020] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] vm_ref = self.build_virtual_machine(instance, [ 725.085020] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 725.085471] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] vif_infos = vmwarevif.get_vif_info(self._session, [ 725.085471] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 725.085471] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] for vif in network_info: [ 725.085471] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 725.085471] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] return self._sync_wrapper(fn, *args, **kwargs) [ 725.085471] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 725.085471] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] self.wait() [ 725.085471] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 725.085471] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] self[:] = self._gt.wait() [ 725.085471] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 725.085471] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] return self._exit_event.wait() [ 725.085471] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 725.085471] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] current.throw(*self._exc) [ 725.085923] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 725.085923] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] result = function(*args, **kwargs) [ 725.085923] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 725.085923] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] return func(*args, **kwargs) [ 725.085923] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 725.085923] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] raise e [ 725.085923] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 725.085923] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] nwinfo = self.network_api.allocate_for_instance( [ 725.085923] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 725.085923] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] created_port_ids = self._update_ports_for_instance( [ 725.085923] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 725.085923] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] with excutils.save_and_reraise_exception(): [ 725.085923] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 725.086369] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] self.force_reraise() [ 725.086369] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 725.086369] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] raise self.value [ 725.086369] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 725.086369] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] updated_port = self._update_port( [ 725.086369] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 725.086369] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] _ensure_no_port_binding_failure(port) [ 725.086369] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 725.086369] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] raise exception.PortBindingFailed(port_id=port['id']) [ 725.086369] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] nova.exception.PortBindingFailed: Binding failed for port 046d1077-d6d9-4ae5-9a33-89a4ba0a7b08, please check neutron logs for more information. [ 725.086369] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] [ 725.088500] env[62965]: INFO nova.compute.manager [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Terminating instance [ 725.271178] env[62965]: INFO nova.compute.manager [-] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Took 1.02 seconds to deallocate network for instance. [ 725.273590] env[62965]: DEBUG nova.compute.claims [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 725.273790] env[62965]: DEBUG oslo_concurrency.lockutils [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.409162] env[62965]: DEBUG oslo_concurrency.lockutils [req-b2996e62-d5f5-4ceb-83be-a998f90b3f49 req-2cdca229-4ca2-4a1b-ad86-59696fb52218 service nova] Releasing lock "refresh_cache-ece118fc-a39e-443a-bfd1-fc332da2452f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.409423] env[62965]: DEBUG nova.compute.manager [req-b2996e62-d5f5-4ceb-83be-a998f90b3f49 req-2cdca229-4ca2-4a1b-ad86-59696fb52218 service nova] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Received event network-vif-deleted-46f93633-3617-45b5-9054-fa6bfc49bee8 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 725.472609] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.470s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.472988] env[62965]: DEBUG nova.compute.manager [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 725.475732] env[62965]: DEBUG oslo_concurrency.lockutils [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 19.590s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.476235] env[62965]: DEBUG nova.objects.instance [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62965) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 725.576615] env[62965]: DEBUG nova.compute.manager [req-a0b0d5f5-563a-4b8e-9c2d-d500db427151 req-799b6b38-0e64-4301-a8de-b7da80262c93 service nova] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Received event network-changed-046d1077-d6d9-4ae5-9a33-89a4ba0a7b08 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 725.576615] env[62965]: DEBUG nova.compute.manager [req-a0b0d5f5-563a-4b8e-9c2d-d500db427151 req-799b6b38-0e64-4301-a8de-b7da80262c93 service nova] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Refreshing instance network info cache due to event network-changed-046d1077-d6d9-4ae5-9a33-89a4ba0a7b08. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 725.576740] env[62965]: DEBUG oslo_concurrency.lockutils [req-a0b0d5f5-563a-4b8e-9c2d-d500db427151 req-799b6b38-0e64-4301-a8de-b7da80262c93 service nova] Acquiring lock "refresh_cache-1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.576940] env[62965]: DEBUG oslo_concurrency.lockutils [req-a0b0d5f5-563a-4b8e-9c2d-d500db427151 req-799b6b38-0e64-4301-a8de-b7da80262c93 service nova] Acquired lock "refresh_cache-1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.577152] env[62965]: DEBUG nova.network.neutron [req-a0b0d5f5-563a-4b8e-9c2d-d500db427151 req-799b6b38-0e64-4301-a8de-b7da80262c93 service nova] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Refreshing network info cache for port 046d1077-d6d9-4ae5-9a33-89a4ba0a7b08 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 725.591383] env[62965]: DEBUG oslo_concurrency.lockutils [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] Acquiring lock "refresh_cache-1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.980347] env[62965]: DEBUG nova.compute.utils [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 725.985118] env[62965]: DEBUG nova.compute.manager [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 725.985172] env[62965]: DEBUG nova.network.neutron [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 726.043325] env[62965]: DEBUG nova.policy [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c9dd0c7ef9bb4a84a126d0da6ca3839d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4c9108a7cd134c92b2781458c899b900', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 726.097319] env[62965]: DEBUG nova.network.neutron [req-a0b0d5f5-563a-4b8e-9c2d-d500db427151 req-799b6b38-0e64-4301-a8de-b7da80262c93 service nova] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.186186] env[62965]: DEBUG nova.network.neutron [req-a0b0d5f5-563a-4b8e-9c2d-d500db427151 req-799b6b38-0e64-4301-a8de-b7da80262c93 service nova] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.414619] env[62965]: DEBUG nova.network.neutron [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Successfully created port: dc4a4c53-a1ac-4fd7-a63e-5a7031939c89 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 726.486031] env[62965]: DEBUG nova.compute.manager [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 726.489331] env[62965]: DEBUG oslo_concurrency.lockutils [None req-df87985c-7060-4181-827c-7193d5e6e55f tempest-ServersAdmin275Test-1319884522 tempest-ServersAdmin275Test-1319884522-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.490513] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.213s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.690763] env[62965]: DEBUG oslo_concurrency.lockutils [req-a0b0d5f5-563a-4b8e-9c2d-d500db427151 req-799b6b38-0e64-4301-a8de-b7da80262c93 service nova] Releasing lock "refresh_cache-1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.691121] env[62965]: DEBUG nova.compute.manager [req-a0b0d5f5-563a-4b8e-9c2d-d500db427151 req-799b6b38-0e64-4301-a8de-b7da80262c93 service nova] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Received event network-vif-deleted-046d1077-d6d9-4ae5-9a33-89a4ba0a7b08 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 726.691511] env[62965]: DEBUG oslo_concurrency.lockutils [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] Acquired lock "refresh_cache-1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.691690] env[62965]: DEBUG nova.network.neutron [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 726.994174] env[62965]: INFO nova.virt.block_device [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Booting with volume 2692164e-9d8f-4fa7-b3c0-b11b8a9a30a5 at /dev/sda [ 727.056916] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fcec32a6-1c77-4aa6-ae3e-95ce1725a901 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.071662] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04ab1fec-e458-4a13-8509-42da07e024a0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.099522] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dca9c618-5d1c-484f-8086-50077e62028d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.107957] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afea7725-1831-46dc-adb6-8b6ed97f8c37 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.135570] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2af3cbe7-e90b-4ee3-a619-215b60d56b7c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.146123] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-759528bf-edfd-4523-b4e8-c77f284760aa {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.159384] env[62965]: DEBUG nova.virt.block_device [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Updating existing volume attachment record: df72c19f-8f5a-479b-8896-d86b620262ff {{(pid=62965) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 727.214061] env[62965]: DEBUG nova.network.neutron [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 727.311556] env[62965]: DEBUG nova.network.neutron [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.384207] env[62965]: DEBUG nova.compute.manager [req-a65ebb74-8a80-4f3d-95e3-2ee1ca4e3088 req-5b7b2149-09aa-4426-be11-749cab166659 service nova] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Received event network-changed-dc4a4c53-a1ac-4fd7-a63e-5a7031939c89 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 727.384414] env[62965]: DEBUG nova.compute.manager [req-a65ebb74-8a80-4f3d-95e3-2ee1ca4e3088 req-5b7b2149-09aa-4426-be11-749cab166659 service nova] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Refreshing instance network info cache due to event network-changed-dc4a4c53-a1ac-4fd7-a63e-5a7031939c89. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 727.384605] env[62965]: DEBUG oslo_concurrency.lockutils [req-a65ebb74-8a80-4f3d-95e3-2ee1ca4e3088 req-5b7b2149-09aa-4426-be11-749cab166659 service nova] Acquiring lock "refresh_cache-99db6bef-6501-4ab3-a329-0877ca303f1b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.384744] env[62965]: DEBUG oslo_concurrency.lockutils [req-a65ebb74-8a80-4f3d-95e3-2ee1ca4e3088 req-5b7b2149-09aa-4426-be11-749cab166659 service nova] Acquired lock "refresh_cache-99db6bef-6501-4ab3-a329-0877ca303f1b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.384897] env[62965]: DEBUG nova.network.neutron [req-a65ebb74-8a80-4f3d-95e3-2ee1ca4e3088 req-5b7b2149-09aa-4426-be11-749cab166659 service nova] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Refreshing network info cache for port dc4a4c53-a1ac-4fd7-a63e-5a7031939c89 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 727.428434] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6dfe849-31e0-47db-8067-0baee647fdcd {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.437265] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8719ff6-9fa3-411a-8a48-243a7e1a4668 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.471824] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd07122d-bb35-42d2-8f60-5d2c2e787dea {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.477114] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18926fa3-494c-4d59-a7ef-333f73f7ee3c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.490866] env[62965]: DEBUG nova.compute.provider_tree [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 727.542280] env[62965]: ERROR nova.compute.manager [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port dc4a4c53-a1ac-4fd7-a63e-5a7031939c89, please check neutron logs for more information. [ 727.542280] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 727.542280] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 727.542280] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 727.542280] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 727.542280] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 727.542280] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 727.542280] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 727.542280] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 727.542280] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 727.542280] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 727.542280] env[62965]: ERROR nova.compute.manager raise self.value [ 727.542280] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 727.542280] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 727.542280] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 727.542280] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 727.543843] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 727.543843] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 727.543843] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port dc4a4c53-a1ac-4fd7-a63e-5a7031939c89, please check neutron logs for more information. [ 727.543843] env[62965]: ERROR nova.compute.manager [ 727.543843] env[62965]: Traceback (most recent call last): [ 727.543843] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 727.543843] env[62965]: listener.cb(fileno) [ 727.543843] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 727.543843] env[62965]: result = function(*args, **kwargs) [ 727.543843] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 727.543843] env[62965]: return func(*args, **kwargs) [ 727.543843] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 727.543843] env[62965]: raise e [ 727.543843] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 727.543843] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 727.543843] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 727.543843] env[62965]: created_port_ids = self._update_ports_for_instance( [ 727.543843] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 727.543843] env[62965]: with excutils.save_and_reraise_exception(): [ 727.543843] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 727.543843] env[62965]: self.force_reraise() [ 727.543843] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 727.543843] env[62965]: raise self.value [ 727.543843] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 727.543843] env[62965]: updated_port = self._update_port( [ 727.543843] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 727.543843] env[62965]: _ensure_no_port_binding_failure(port) [ 727.543843] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 727.543843] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 727.545795] env[62965]: nova.exception.PortBindingFailed: Binding failed for port dc4a4c53-a1ac-4fd7-a63e-5a7031939c89, please check neutron logs for more information. [ 727.545795] env[62965]: Removing descriptor: 17 [ 727.814407] env[62965]: DEBUG oslo_concurrency.lockutils [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] Releasing lock "refresh_cache-1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.814834] env[62965]: DEBUG nova.compute.manager [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 727.815039] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 727.815321] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4248e612-0402-4e9e-b3f5-c8b843c7717e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.825253] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b00f9acb-b45d-4784-af9b-c34d39c3abd7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.846519] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb could not be found. [ 727.846675] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 727.846860] env[62965]: INFO nova.compute.manager [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Took 0.03 seconds to destroy the instance on the hypervisor. [ 727.847161] env[62965]: DEBUG oslo.service.loopingcall [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 727.847385] env[62965]: DEBUG nova.compute.manager [-] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 727.847481] env[62965]: DEBUG nova.network.neutron [-] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 727.862841] env[62965]: DEBUG nova.network.neutron [-] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 727.904881] env[62965]: DEBUG nova.network.neutron [req-a65ebb74-8a80-4f3d-95e3-2ee1ca4e3088 req-5b7b2149-09aa-4426-be11-749cab166659 service nova] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 727.995704] env[62965]: DEBUG nova.scheduler.client.report [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 728.000822] env[62965]: DEBUG nova.network.neutron [req-a65ebb74-8a80-4f3d-95e3-2ee1ca4e3088 req-5b7b2149-09aa-4426-be11-749cab166659 service nova] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.365410] env[62965]: DEBUG nova.network.neutron [-] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.502038] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.011s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.502550] env[62965]: ERROR nova.compute.manager [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 42ad3568-1e88-47f0-bf93-9f3dfe2266c8, please check neutron logs for more information. [ 728.502550] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Traceback (most recent call last): [ 728.502550] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 728.502550] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] self.driver.spawn(context, instance, image_meta, [ 728.502550] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 728.502550] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] self._vmops.spawn(context, instance, image_meta, injected_files, [ 728.502550] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 728.502550] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] vm_ref = self.build_virtual_machine(instance, [ 728.502550] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 728.502550] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] vif_infos = vmwarevif.get_vif_info(self._session, [ 728.502550] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 728.502943] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] for vif in network_info: [ 728.502943] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 728.502943] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] return self._sync_wrapper(fn, *args, **kwargs) [ 728.502943] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 728.502943] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] self.wait() [ 728.502943] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 728.502943] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] self[:] = self._gt.wait() [ 728.502943] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 728.502943] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] return self._exit_event.wait() [ 728.502943] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 728.502943] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] result = hub.switch() [ 728.502943] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 728.502943] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] return self.greenlet.switch() [ 728.503394] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 728.503394] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] result = function(*args, **kwargs) [ 728.503394] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 728.503394] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] return func(*args, **kwargs) [ 728.503394] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 728.503394] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] raise e [ 728.503394] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 728.503394] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] nwinfo = self.network_api.allocate_for_instance( [ 728.503394] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 728.503394] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] created_port_ids = self._update_ports_for_instance( [ 728.503394] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 728.503394] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] with excutils.save_and_reraise_exception(): [ 728.503394] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.503770] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] self.force_reraise() [ 728.503770] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.503770] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] raise self.value [ 728.503770] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 728.503770] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] updated_port = self._update_port( [ 728.503770] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.503770] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] _ensure_no_port_binding_failure(port) [ 728.503770] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.503770] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] raise exception.PortBindingFailed(port_id=port['id']) [ 728.503770] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] nova.exception.PortBindingFailed: Binding failed for port 42ad3568-1e88-47f0-bf93-9f3dfe2266c8, please check neutron logs for more information. [ 728.503770] env[62965]: ERROR nova.compute.manager [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] [ 728.504157] env[62965]: DEBUG nova.compute.utils [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Binding failed for port 42ad3568-1e88-47f0-bf93-9f3dfe2266c8, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 728.504706] env[62965]: DEBUG oslo_concurrency.lockutils [req-a65ebb74-8a80-4f3d-95e3-2ee1ca4e3088 req-5b7b2149-09aa-4426-be11-749cab166659 service nova] Releasing lock "refresh_cache-99db6bef-6501-4ab3-a329-0877ca303f1b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.505251] env[62965]: DEBUG nova.compute.manager [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Build of instance 4519d37a-f8e6-4a25-99a1-399925199a58 was re-scheduled: Binding failed for port 42ad3568-1e88-47f0-bf93-9f3dfe2266c8, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 728.505661] env[62965]: DEBUG nova.compute.manager [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 728.505947] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Acquiring lock "refresh_cache-4519d37a-f8e6-4a25-99a1-399925199a58" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.506105] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Acquired lock "refresh_cache-4519d37a-f8e6-4a25-99a1-399925199a58" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.506266] env[62965]: DEBUG nova.network.neutron [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 728.507209] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.641s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.508655] env[62965]: INFO nova.compute.claims [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 728.868054] env[62965]: INFO nova.compute.manager [-] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Took 1.02 seconds to deallocate network for instance. [ 728.870354] env[62965]: DEBUG nova.compute.claims [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 728.870602] env[62965]: DEBUG oslo_concurrency.lockutils [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.044384] env[62965]: DEBUG nova.network.neutron [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 729.116804] env[62965]: DEBUG nova.network.neutron [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.265217] env[62965]: DEBUG nova.compute.manager [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 729.265752] env[62965]: DEBUG nova.virt.hardware [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 729.265962] env[62965]: DEBUG nova.virt.hardware [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 729.266130] env[62965]: DEBUG nova.virt.hardware [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 729.266308] env[62965]: DEBUG nova.virt.hardware [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 729.266453] env[62965]: DEBUG nova.virt.hardware [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 729.266632] env[62965]: DEBUG nova.virt.hardware [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 729.266837] env[62965]: DEBUG nova.virt.hardware [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 729.266993] env[62965]: DEBUG nova.virt.hardware [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 729.267172] env[62965]: DEBUG nova.virt.hardware [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 729.267331] env[62965]: DEBUG nova.virt.hardware [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 729.267498] env[62965]: DEBUG nova.virt.hardware [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 729.268353] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efa7e7f9-fb5c-4878-9c49-370d6d7d025d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.276689] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aaf22b4-9b8f-4308-b6de-6dd0ee71227b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.290321] env[62965]: ERROR nova.compute.manager [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port dc4a4c53-a1ac-4fd7-a63e-5a7031939c89, please check neutron logs for more information. [ 729.290321] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Traceback (most recent call last): [ 729.290321] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 729.290321] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] yield resources [ 729.290321] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 729.290321] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] self.driver.spawn(context, instance, image_meta, [ 729.290321] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 729.290321] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 729.290321] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 729.290321] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] vm_ref = self.build_virtual_machine(instance, [ 729.290321] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 729.290830] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] vif_infos = vmwarevif.get_vif_info(self._session, [ 729.290830] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 729.290830] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] for vif in network_info: [ 729.290830] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 729.290830] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] return self._sync_wrapper(fn, *args, **kwargs) [ 729.290830] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 729.290830] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] self.wait() [ 729.290830] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 729.290830] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] self[:] = self._gt.wait() [ 729.290830] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 729.290830] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] return self._exit_event.wait() [ 729.290830] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 729.290830] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] current.throw(*self._exc) [ 729.291414] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 729.291414] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] result = function(*args, **kwargs) [ 729.291414] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 729.291414] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] return func(*args, **kwargs) [ 729.291414] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 729.291414] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] raise e [ 729.291414] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 729.291414] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] nwinfo = self.network_api.allocate_for_instance( [ 729.291414] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 729.291414] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] created_port_ids = self._update_ports_for_instance( [ 729.291414] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 729.291414] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] with excutils.save_and_reraise_exception(): [ 729.291414] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 729.291840] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] self.force_reraise() [ 729.291840] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 729.291840] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] raise self.value [ 729.291840] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 729.291840] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] updated_port = self._update_port( [ 729.291840] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 729.291840] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] _ensure_no_port_binding_failure(port) [ 729.291840] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 729.291840] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] raise exception.PortBindingFailed(port_id=port['id']) [ 729.291840] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] nova.exception.PortBindingFailed: Binding failed for port dc4a4c53-a1ac-4fd7-a63e-5a7031939c89, please check neutron logs for more information. [ 729.291840] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] [ 729.291840] env[62965]: INFO nova.compute.manager [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Terminating instance [ 729.412156] env[62965]: DEBUG nova.compute.manager [req-cd844304-64b6-4b12-a724-8c2830ac7b76 req-1ddb5781-af51-4ac2-9b1a-cbe7373c915f service nova] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Received event network-vif-deleted-dc4a4c53-a1ac-4fd7-a63e-5a7031939c89 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 729.620080] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Releasing lock "refresh_cache-4519d37a-f8e6-4a25-99a1-399925199a58" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.620326] env[62965]: DEBUG nova.compute.manager [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 729.620486] env[62965]: DEBUG nova.compute.manager [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 729.620650] env[62965]: DEBUG nova.network.neutron [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 729.647419] env[62965]: DEBUG nova.network.neutron [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 729.794194] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] Acquiring lock "refresh_cache-99db6bef-6501-4ab3-a329-0877ca303f1b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.794373] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] Acquired lock "refresh_cache-99db6bef-6501-4ab3-a329-0877ca303f1b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.794544] env[62965]: DEBUG nova.network.neutron [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 729.833927] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-504c40e4-dad3-4b94-a0e3-05fab6013e65 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.841748] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33170dea-7838-4dc1-ad25-ceebfad740bf {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.871397] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-125a83c0-6477-497a-a8fb-723569193f60 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.878426] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f5eae20-f59d-47f2-8f00-b3d6028da784 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.891142] env[62965]: DEBUG nova.compute.provider_tree [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 730.151059] env[62965]: DEBUG nova.network.neutron [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.324952] env[62965]: DEBUG nova.network.neutron [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 730.394073] env[62965]: DEBUG nova.scheduler.client.report [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 730.401421] env[62965]: DEBUG nova.network.neutron [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.653273] env[62965]: INFO nova.compute.manager [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: 4519d37a-f8e6-4a25-99a1-399925199a58] Took 1.03 seconds to deallocate network for instance. [ 730.899922] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.393s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.900457] env[62965]: DEBUG nova.compute.manager [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 730.903436] env[62965]: DEBUG oslo_concurrency.lockutils [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.777s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.903436] env[62965]: DEBUG nova.objects.instance [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Lazy-loading 'resources' on Instance uuid 02f113d2-f991-4268-b9a4-59e2e6bcf7a6 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 730.906393] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] Releasing lock "refresh_cache-99db6bef-6501-4ab3-a329-0877ca303f1b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.906393] env[62965]: DEBUG nova.compute.manager [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 730.906393] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b61f8aa4-1be7-4b79-98b5-2c3ab52cf59f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.915633] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8fad99e-83b8-4508-8285-0aa2e33d1dad {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.938674] env[62965]: WARNING nova.virt.vmwareapi.driver [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 99db6bef-6501-4ab3-a329-0877ca303f1b could not be found. [ 730.938987] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 730.939325] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b6854faa-1019-4f09-a265-768c78e1d700 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.947537] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06bd9305-006f-408b-b4c1-bbf4e200f386 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.970758] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 99db6bef-6501-4ab3-a329-0877ca303f1b could not be found. [ 730.970970] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 730.971171] env[62965]: INFO nova.compute.manager [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Took 0.07 seconds to destroy the instance on the hypervisor. [ 730.971411] env[62965]: DEBUG oslo.service.loopingcall [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 730.971630] env[62965]: DEBUG nova.compute.manager [-] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 730.971725] env[62965]: DEBUG nova.network.neutron [-] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 730.986090] env[62965]: DEBUG nova.network.neutron [-] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 731.408988] env[62965]: DEBUG nova.compute.utils [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 731.412953] env[62965]: DEBUG nova.compute.manager [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 731.412953] env[62965]: DEBUG nova.network.neutron [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 731.456855] env[62965]: DEBUG nova.policy [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7e7bde3eb8774dd2927e9a4ea9dd4cf4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '759a695bb7ee41a78d38ebf69e6bd51f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 731.488082] env[62965]: DEBUG nova.network.neutron [-] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.682242] env[62965]: INFO nova.scheduler.client.report [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Deleted allocations for instance 4519d37a-f8e6-4a25-99a1-399925199a58 [ 731.769881] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f8957db-21cf-4538-8711-ce896d197ba4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.778894] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56209073-412a-4732-abcb-0c0a7e7f82f1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.808471] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b4959f9-325c-48e6-85a1-9a545830a77c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.815405] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dca29d81-94bd-4e4e-af62-d700e0925f90 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.827980] env[62965]: DEBUG nova.compute.provider_tree [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 731.913319] env[62965]: DEBUG nova.compute.manager [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 731.986300] env[62965]: DEBUG nova.network.neutron [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Successfully created port: 7af92ac6-9925-4abe-a323-98c576e48b87 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 731.992020] env[62965]: INFO nova.compute.manager [-] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Took 1.02 seconds to deallocate network for instance. [ 732.190177] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e2530cd1-1301-4c15-9621-47f26ccd179b tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Lock "4519d37a-f8e6-4a25-99a1-399925199a58" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 149.048s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.332664] env[62965]: DEBUG nova.scheduler.client.report [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 732.553996] env[62965]: INFO nova.compute.manager [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Took 0.56 seconds to detach 1 volumes for instance. [ 732.556711] env[62965]: DEBUG nova.compute.claims [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 732.556711] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 732.693923] env[62965]: DEBUG nova.compute.manager [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 732.838385] env[62965]: DEBUG oslo_concurrency.lockutils [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.935s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.841730] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.743s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.866184] env[62965]: INFO nova.scheduler.client.report [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Deleted allocations for instance 02f113d2-f991-4268-b9a4-59e2e6bcf7a6 [ 732.926793] env[62965]: DEBUG nova.compute.manager [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 732.947577] env[62965]: DEBUG nova.virt.hardware [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 732.947816] env[62965]: DEBUG nova.virt.hardware [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 732.947969] env[62965]: DEBUG nova.virt.hardware [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 732.948380] env[62965]: DEBUG nova.virt.hardware [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 732.948713] env[62965]: DEBUG nova.virt.hardware [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 732.949811] env[62965]: DEBUG nova.virt.hardware [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 732.950106] env[62965]: DEBUG nova.virt.hardware [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 732.950278] env[62965]: DEBUG nova.virt.hardware [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 732.950536] env[62965]: DEBUG nova.virt.hardware [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 732.950733] env[62965]: DEBUG nova.virt.hardware [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 732.950911] env[62965]: DEBUG nova.virt.hardware [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 732.952083] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c0ffe4d-36a5-4070-a800-1d827480611d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.961903] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ce2aae2-99ec-45df-80ec-da20b66ada4f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.045489] env[62965]: DEBUG nova.compute.manager [req-a92cfb8a-12b6-4104-9d3b-1fdb4f72e30f req-471b403f-7b45-44db-a449-cbc4c880af1a service nova] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Received event network-changed-7af92ac6-9925-4abe-a323-98c576e48b87 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 733.045689] env[62965]: DEBUG nova.compute.manager [req-a92cfb8a-12b6-4104-9d3b-1fdb4f72e30f req-471b403f-7b45-44db-a449-cbc4c880af1a service nova] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Refreshing instance network info cache due to event network-changed-7af92ac6-9925-4abe-a323-98c576e48b87. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 733.045901] env[62965]: DEBUG oslo_concurrency.lockutils [req-a92cfb8a-12b6-4104-9d3b-1fdb4f72e30f req-471b403f-7b45-44db-a449-cbc4c880af1a service nova] Acquiring lock "refresh_cache-8342b9df-ebfd-403b-a0b8-1fdd6ebe2add" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.046049] env[62965]: DEBUG oslo_concurrency.lockutils [req-a92cfb8a-12b6-4104-9d3b-1fdb4f72e30f req-471b403f-7b45-44db-a449-cbc4c880af1a service nova] Acquired lock "refresh_cache-8342b9df-ebfd-403b-a0b8-1fdd6ebe2add" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.046205] env[62965]: DEBUG nova.network.neutron [req-a92cfb8a-12b6-4104-9d3b-1fdb4f72e30f req-471b403f-7b45-44db-a449-cbc4c880af1a service nova] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Refreshing network info cache for port 7af92ac6-9925-4abe-a323-98c576e48b87 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 733.224017] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.270193] env[62965]: ERROR nova.compute.manager [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7af92ac6-9925-4abe-a323-98c576e48b87, please check neutron logs for more information. [ 733.270193] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 733.270193] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 733.270193] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 733.270193] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 733.270193] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 733.270193] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 733.270193] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 733.270193] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.270193] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 733.270193] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.270193] env[62965]: ERROR nova.compute.manager raise self.value [ 733.270193] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 733.270193] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 733.270193] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 733.270193] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 733.270791] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 733.270791] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 733.270791] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7af92ac6-9925-4abe-a323-98c576e48b87, please check neutron logs for more information. [ 733.270791] env[62965]: ERROR nova.compute.manager [ 733.270791] env[62965]: Traceback (most recent call last): [ 733.270791] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 733.270791] env[62965]: listener.cb(fileno) [ 733.270791] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 733.270791] env[62965]: result = function(*args, **kwargs) [ 733.270791] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 733.270791] env[62965]: return func(*args, **kwargs) [ 733.270791] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 733.270791] env[62965]: raise e [ 733.270791] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 733.270791] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 733.270791] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 733.270791] env[62965]: created_port_ids = self._update_ports_for_instance( [ 733.270791] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 733.270791] env[62965]: with excutils.save_and_reraise_exception(): [ 733.270791] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.270791] env[62965]: self.force_reraise() [ 733.270791] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.270791] env[62965]: raise self.value [ 733.270791] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 733.270791] env[62965]: updated_port = self._update_port( [ 733.270791] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 733.270791] env[62965]: _ensure_no_port_binding_failure(port) [ 733.270791] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 733.270791] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 733.271685] env[62965]: nova.exception.PortBindingFailed: Binding failed for port 7af92ac6-9925-4abe-a323-98c576e48b87, please check neutron logs for more information. [ 733.271685] env[62965]: Removing descriptor: 17 [ 733.271685] env[62965]: ERROR nova.compute.manager [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7af92ac6-9925-4abe-a323-98c576e48b87, please check neutron logs for more information. [ 733.271685] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Traceback (most recent call last): [ 733.271685] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 733.271685] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] yield resources [ 733.271685] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 733.271685] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] self.driver.spawn(context, instance, image_meta, [ 733.271685] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 733.271685] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] self._vmops.spawn(context, instance, image_meta, injected_files, [ 733.271685] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 733.271685] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] vm_ref = self.build_virtual_machine(instance, [ 733.272035] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 733.272035] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] vif_infos = vmwarevif.get_vif_info(self._session, [ 733.272035] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 733.272035] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] for vif in network_info: [ 733.272035] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 733.272035] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] return self._sync_wrapper(fn, *args, **kwargs) [ 733.272035] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 733.272035] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] self.wait() [ 733.272035] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 733.272035] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] self[:] = self._gt.wait() [ 733.272035] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 733.272035] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] return self._exit_event.wait() [ 733.272035] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 733.272433] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] result = hub.switch() [ 733.272433] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 733.272433] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] return self.greenlet.switch() [ 733.272433] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 733.272433] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] result = function(*args, **kwargs) [ 733.272433] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 733.272433] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] return func(*args, **kwargs) [ 733.272433] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 733.272433] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] raise e [ 733.272433] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 733.272433] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] nwinfo = self.network_api.allocate_for_instance( [ 733.272433] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 733.272433] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] created_port_ids = self._update_ports_for_instance( [ 733.272865] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 733.272865] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] with excutils.save_and_reraise_exception(): [ 733.272865] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.272865] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] self.force_reraise() [ 733.272865] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.272865] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] raise self.value [ 733.272865] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 733.272865] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] updated_port = self._update_port( [ 733.272865] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 733.272865] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] _ensure_no_port_binding_failure(port) [ 733.272865] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 733.272865] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] raise exception.PortBindingFailed(port_id=port['id']) [ 733.273287] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] nova.exception.PortBindingFailed: Binding failed for port 7af92ac6-9925-4abe-a323-98c576e48b87, please check neutron logs for more information. [ 733.273287] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] [ 733.273287] env[62965]: INFO nova.compute.manager [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Terminating instance [ 733.376843] env[62965]: DEBUG oslo_concurrency.lockutils [None req-232e19cc-08b3-407e-9608-5426fb359e61 tempest-ServersAdmin275Test-117833045 tempest-ServersAdmin275Test-117833045-project-member] Lock "02f113d2-f991-4268-b9a4-59e2e6bcf7a6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.465s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.590039] env[62965]: DEBUG nova.network.neutron [req-a92cfb8a-12b6-4104-9d3b-1fdb4f72e30f req-471b403f-7b45-44db-a449-cbc4c880af1a service nova] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 733.650709] env[62965]: DEBUG oslo_concurrency.lockutils [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Acquiring lock "e9cac9cf-c165-428b-9d83-e37905f42c93" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.650937] env[62965]: DEBUG oslo_concurrency.lockutils [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Lock "e9cac9cf-c165-428b-9d83-e37905f42c93" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.718716] env[62965]: DEBUG nova.network.neutron [req-a92cfb8a-12b6-4104-9d3b-1fdb4f72e30f req-471b403f-7b45-44db-a449-cbc4c880af1a service nova] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.725639] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b201db1-6dec-4121-9b9b-754239637722 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.733457] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7814ee01-2434-446f-8034-4e723f61d569 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.765310] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af161c98-539a-4c9b-bd78-4f718e8c985b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.773992] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fee2f29-c3a5-464d-83b3-21d6fa9e452d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.779164] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] Acquiring lock "refresh_cache-8342b9df-ebfd-403b-a0b8-1fdd6ebe2add" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.790656] env[62965]: DEBUG nova.compute.provider_tree [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 734.224722] env[62965]: DEBUG oslo_concurrency.lockutils [req-a92cfb8a-12b6-4104-9d3b-1fdb4f72e30f req-471b403f-7b45-44db-a449-cbc4c880af1a service nova] Releasing lock "refresh_cache-8342b9df-ebfd-403b-a0b8-1fdd6ebe2add" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.225603] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] Acquired lock "refresh_cache-8342b9df-ebfd-403b-a0b8-1fdd6ebe2add" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.225875] env[62965]: DEBUG nova.network.neutron [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 734.294374] env[62965]: DEBUG nova.scheduler.client.report [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 734.744042] env[62965]: DEBUG nova.network.neutron [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 734.798936] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.958s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 734.799680] env[62965]: ERROR nova.compute.manager [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port cfb0fc29-145a-45a9-8e88-4fb6176ed1e4, please check neutron logs for more information. [ 734.799680] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Traceback (most recent call last): [ 734.799680] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 734.799680] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] self.driver.spawn(context, instance, image_meta, [ 734.799680] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 734.799680] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] self._vmops.spawn(context, instance, image_meta, injected_files, [ 734.799680] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 734.799680] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] vm_ref = self.build_virtual_machine(instance, [ 734.799680] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 734.799680] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] vif_infos = vmwarevif.get_vif_info(self._session, [ 734.799680] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 734.800339] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] for vif in network_info: [ 734.800339] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 734.800339] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] return self._sync_wrapper(fn, *args, **kwargs) [ 734.800339] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 734.800339] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] self.wait() [ 734.800339] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 734.800339] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] self[:] = self._gt.wait() [ 734.800339] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 734.800339] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] return self._exit_event.wait() [ 734.800339] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 734.800339] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] result = hub.switch() [ 734.800339] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 734.800339] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] return self.greenlet.switch() [ 734.800811] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 734.800811] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] result = function(*args, **kwargs) [ 734.800811] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 734.800811] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] return func(*args, **kwargs) [ 734.800811] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 734.800811] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] raise e [ 734.800811] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 734.800811] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] nwinfo = self.network_api.allocate_for_instance( [ 734.800811] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 734.800811] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] created_port_ids = self._update_ports_for_instance( [ 734.800811] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 734.800811] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] with excutils.save_and_reraise_exception(): [ 734.800811] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 734.801290] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] self.force_reraise() [ 734.801290] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 734.801290] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] raise self.value [ 734.801290] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 734.801290] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] updated_port = self._update_port( [ 734.801290] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 734.801290] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] _ensure_no_port_binding_failure(port) [ 734.801290] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 734.801290] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] raise exception.PortBindingFailed(port_id=port['id']) [ 734.801290] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] nova.exception.PortBindingFailed: Binding failed for port cfb0fc29-145a-45a9-8e88-4fb6176ed1e4, please check neutron logs for more information. [ 734.801290] env[62965]: ERROR nova.compute.manager [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] [ 734.802188] env[62965]: DEBUG nova.compute.utils [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Binding failed for port cfb0fc29-145a-45a9-8e88-4fb6176ed1e4, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 734.804316] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.741s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 734.809274] env[62965]: DEBUG nova.compute.manager [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Build of instance 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee was re-scheduled: Binding failed for port cfb0fc29-145a-45a9-8e88-4fb6176ed1e4, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 734.809274] env[62965]: DEBUG nova.compute.manager [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 734.809274] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] Acquiring lock "refresh_cache-6c7ed2a3-ed0f-476b-8232-d7a3c27adcee" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.809274] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] Acquired lock "refresh_cache-6c7ed2a3-ed0f-476b-8232-d7a3c27adcee" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.809515] env[62965]: DEBUG nova.network.neutron [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 734.869578] env[62965]: DEBUG nova.network.neutron [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.068368] env[62965]: DEBUG nova.compute.manager [req-2c20e9d3-6b28-458d-9f1b-e99498e09615 req-ab87c4e1-28d4-438c-9ad5-243f57e042f0 service nova] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Received event network-vif-deleted-7af92ac6-9925-4abe-a323-98c576e48b87 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 735.329905] env[62965]: DEBUG nova.network.neutron [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.372515] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] Releasing lock "refresh_cache-8342b9df-ebfd-403b-a0b8-1fdd6ebe2add" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.373837] env[62965]: DEBUG nova.compute.manager [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 735.373837] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 735.373837] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-25906c6a-824c-4e58-8326-4a5c6b801b43 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.382489] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13541827-6e9e-41f7-827d-40754b1b5046 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.407947] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add could not be found. [ 735.411083] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 735.411295] env[62965]: INFO nova.compute.manager [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Took 0.04 seconds to destroy the instance on the hypervisor. [ 735.411546] env[62965]: DEBUG oslo.service.loopingcall [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 735.416020] env[62965]: DEBUG nova.network.neutron [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.416020] env[62965]: DEBUG nova.compute.manager [-] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 735.416020] env[62965]: DEBUG nova.network.neutron [-] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 735.441100] env[62965]: DEBUG nova.network.neutron [-] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.719764] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e73ac20-f8f5-4536-9561-720e35ea936e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.727413] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-611f0f8f-a4de-4e8a-91d8-ebdcf8c1ff2e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.764322] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-783cf0e1-4e73-49c8-bd4b-94347861c355 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.773075] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f1c2a9e-8eba-48c6-8e76-b1de2c016104 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.789594] env[62965]: DEBUG nova.compute.provider_tree [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 735.919783] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] Releasing lock "refresh_cache-6c7ed2a3-ed0f-476b-8232-d7a3c27adcee" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.920147] env[62965]: DEBUG nova.compute.manager [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 735.920391] env[62965]: DEBUG nova.compute.manager [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 735.920736] env[62965]: DEBUG nova.network.neutron [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 735.937921] env[62965]: DEBUG nova.network.neutron [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.943483] env[62965]: DEBUG nova.network.neutron [-] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.297818] env[62965]: DEBUG nova.scheduler.client.report [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 736.441217] env[62965]: DEBUG nova.network.neutron [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.446085] env[62965]: INFO nova.compute.manager [-] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Took 1.03 seconds to deallocate network for instance. [ 736.449416] env[62965]: DEBUG nova.compute.claims [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 736.449416] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.807028] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.003s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.807604] env[62965]: ERROR nova.compute.manager [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1bd4d220-93d4-492d-9b69-965c342a2559, please check neutron logs for more information. [ 736.807604] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Traceback (most recent call last): [ 736.807604] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 736.807604] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] self.driver.spawn(context, instance, image_meta, [ 736.807604] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 736.807604] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] self._vmops.spawn(context, instance, image_meta, injected_files, [ 736.807604] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 736.807604] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] vm_ref = self.build_virtual_machine(instance, [ 736.807604] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 736.807604] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] vif_infos = vmwarevif.get_vif_info(self._session, [ 736.807604] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 736.807946] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] for vif in network_info: [ 736.807946] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 736.807946] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] return self._sync_wrapper(fn, *args, **kwargs) [ 736.807946] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 736.807946] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] self.wait() [ 736.807946] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 736.807946] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] self[:] = self._gt.wait() [ 736.807946] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 736.807946] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] return self._exit_event.wait() [ 736.807946] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 736.807946] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] current.throw(*self._exc) [ 736.807946] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 736.807946] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] result = function(*args, **kwargs) [ 736.808320] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 736.808320] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] return func(*args, **kwargs) [ 736.808320] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 736.808320] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] raise e [ 736.808320] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 736.808320] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] nwinfo = self.network_api.allocate_for_instance( [ 736.808320] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 736.808320] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] created_port_ids = self._update_ports_for_instance( [ 736.808320] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 736.808320] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] with excutils.save_and_reraise_exception(): [ 736.808320] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 736.808320] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] self.force_reraise() [ 736.808320] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 736.808795] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] raise self.value [ 736.808795] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 736.808795] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] updated_port = self._update_port( [ 736.808795] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 736.808795] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] _ensure_no_port_binding_failure(port) [ 736.808795] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 736.808795] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] raise exception.PortBindingFailed(port_id=port['id']) [ 736.808795] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] nova.exception.PortBindingFailed: Binding failed for port 1bd4d220-93d4-492d-9b69-965c342a2559, please check neutron logs for more information. [ 736.808795] env[62965]: ERROR nova.compute.manager [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] [ 736.808795] env[62965]: DEBUG nova.compute.utils [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Binding failed for port 1bd4d220-93d4-492d-9b69-965c342a2559, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 736.809810] env[62965]: DEBUG oslo_concurrency.lockutils [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.125s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.815915] env[62965]: DEBUG nova.compute.manager [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Build of instance 852e7cdf-5da3-4217-8960-f3204d4c1035 was re-scheduled: Binding failed for port 1bd4d220-93d4-492d-9b69-965c342a2559, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 736.816390] env[62965]: DEBUG nova.compute.manager [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 736.816601] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] Acquiring lock "refresh_cache-852e7cdf-5da3-4217-8960-f3204d4c1035" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.816755] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] Acquired lock "refresh_cache-852e7cdf-5da3-4217-8960-f3204d4c1035" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.817010] env[62965]: DEBUG nova.network.neutron [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 736.948977] env[62965]: INFO nova.compute.manager [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] [instance: 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee] Took 1.02 seconds to deallocate network for instance. [ 737.353109] env[62965]: DEBUG nova.network.neutron [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 737.478540] env[62965]: DEBUG nova.network.neutron [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.695170] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ebdb30f-af80-4492-add3-f387df1f52e3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.702202] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22db3a33-b39b-40ed-9bb8-599e7d92a827 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.732353] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd45e0ec-139a-46ff-a8a6-fbf86ffa7e83 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.740511] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c64e1e4-682f-4808-acf1-a6aa098baa47 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.754719] env[62965]: DEBUG nova.compute.provider_tree [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 737.975333] env[62965]: INFO nova.scheduler.client.report [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] Deleted allocations for instance 6c7ed2a3-ed0f-476b-8232-d7a3c27adcee [ 737.987023] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] Releasing lock "refresh_cache-852e7cdf-5da3-4217-8960-f3204d4c1035" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.987023] env[62965]: DEBUG nova.compute.manager [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 737.987023] env[62965]: DEBUG nova.compute.manager [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 737.987023] env[62965]: DEBUG nova.network.neutron [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 738.022154] env[62965]: DEBUG nova.network.neutron [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.111701] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] Acquiring lock "3d0f74d4-ff59-462c-9577-c01c4f15ea1d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.111826] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] Lock "3d0f74d4-ff59-462c-9577-c01c4f15ea1d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.258438] env[62965]: DEBUG nova.scheduler.client.report [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 738.488016] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a5eee249-92a4-401d-8f4e-0cd00327d771 tempest-ServerDiagnosticsNegativeTest-1535587098 tempest-ServerDiagnosticsNegativeTest-1535587098-project-member] Lock "6c7ed2a3-ed0f-476b-8232-d7a3c27adcee" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 153.816s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.525604] env[62965]: DEBUG nova.network.neutron [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.763944] env[62965]: DEBUG oslo_concurrency.lockutils [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.954s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.765563] env[62965]: ERROR nova.compute.manager [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 832ab8bd-d991-46d2-b86c-20147968b4fe, please check neutron logs for more information. [ 738.765563] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Traceback (most recent call last): [ 738.765563] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 738.765563] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] self.driver.spawn(context, instance, image_meta, [ 738.765563] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 738.765563] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] self._vmops.spawn(context, instance, image_meta, injected_files, [ 738.765563] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 738.765563] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] vm_ref = self.build_virtual_machine(instance, [ 738.765563] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 738.765563] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] vif_infos = vmwarevif.get_vif_info(self._session, [ 738.765563] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 738.766089] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] for vif in network_info: [ 738.766089] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 738.766089] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] return self._sync_wrapper(fn, *args, **kwargs) [ 738.766089] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 738.766089] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] self.wait() [ 738.766089] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 738.766089] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] self[:] = self._gt.wait() [ 738.766089] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 738.766089] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] return self._exit_event.wait() [ 738.766089] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 738.766089] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] result = hub.switch() [ 738.766089] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 738.766089] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] return self.greenlet.switch() [ 738.766513] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 738.766513] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] result = function(*args, **kwargs) [ 738.766513] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 738.766513] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] return func(*args, **kwargs) [ 738.766513] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 738.766513] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] raise e [ 738.766513] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 738.766513] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] nwinfo = self.network_api.allocate_for_instance( [ 738.766513] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 738.766513] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] created_port_ids = self._update_ports_for_instance( [ 738.766513] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 738.766513] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] with excutils.save_and_reraise_exception(): [ 738.766513] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.766954] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] self.force_reraise() [ 738.766954] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.766954] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] raise self.value [ 738.766954] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 738.766954] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] updated_port = self._update_port( [ 738.766954] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.766954] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] _ensure_no_port_binding_failure(port) [ 738.766954] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.766954] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] raise exception.PortBindingFailed(port_id=port['id']) [ 738.766954] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] nova.exception.PortBindingFailed: Binding failed for port 832ab8bd-d991-46d2-b86c-20147968b4fe, please check neutron logs for more information. [ 738.766954] env[62965]: ERROR nova.compute.manager [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] [ 738.767428] env[62965]: DEBUG nova.compute.utils [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Binding failed for port 832ab8bd-d991-46d2-b86c-20147968b4fe, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 738.767428] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.461s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.769757] env[62965]: DEBUG nova.compute.manager [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Build of instance b2f38e92-37a1-4e0b-9292-b20ea43a2038 was re-scheduled: Binding failed for port 832ab8bd-d991-46d2-b86c-20147968b4fe, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 738.770203] env[62965]: DEBUG nova.compute.manager [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 738.770430] env[62965]: DEBUG oslo_concurrency.lockutils [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] Acquiring lock "refresh_cache-b2f38e92-37a1-4e0b-9292-b20ea43a2038" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.770576] env[62965]: DEBUG oslo_concurrency.lockutils [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] Acquired lock "refresh_cache-b2f38e92-37a1-4e0b-9292-b20ea43a2038" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.770731] env[62965]: DEBUG nova.network.neutron [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 738.991247] env[62965]: DEBUG nova.compute.manager [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 739.028873] env[62965]: INFO nova.compute.manager [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] [instance: 852e7cdf-5da3-4217-8960-f3204d4c1035] Took 1.04 seconds to deallocate network for instance. [ 739.295333] env[62965]: DEBUG nova.network.neutron [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.407449] env[62965]: DEBUG nova.network.neutron [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.506798] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 739.578120] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-042b35e2-ab2a-45d2-a570-2e5fe9e63c1f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.587033] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc1693a9-e07f-4063-88e6-c99352f1b467 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.617175] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6539a2e-5cd7-424e-b174-8f4e7ed4d79e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.624413] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16930790-97c7-40f3-a284-d7a8215a91c5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.637353] env[62965]: DEBUG nova.compute.provider_tree [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 739.910587] env[62965]: DEBUG oslo_concurrency.lockutils [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] Releasing lock "refresh_cache-b2f38e92-37a1-4e0b-9292-b20ea43a2038" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.910789] env[62965]: DEBUG nova.compute.manager [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 739.910974] env[62965]: DEBUG nova.compute.manager [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 739.911182] env[62965]: DEBUG nova.network.neutron [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 739.929199] env[62965]: DEBUG nova.network.neutron [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.053763] env[62965]: INFO nova.scheduler.client.report [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] Deleted allocations for instance 852e7cdf-5da3-4217-8960-f3204d4c1035 [ 740.140623] env[62965]: DEBUG nova.scheduler.client.report [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 740.433040] env[62965]: DEBUG nova.network.neutron [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.561772] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f17d0269-64cc-4772-941f-4d7b37a6b34c tempest-ServerPasswordTestJSON-1870144464 tempest-ServerPasswordTestJSON-1870144464-project-member] Lock "852e7cdf-5da3-4217-8960-f3204d4c1035" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 154.160s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.645788] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.879s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.646409] env[62965]: ERROR nova.compute.manager [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 84318521-767d-49b3-81f5-02cbb8556c81, please check neutron logs for more information. [ 740.646409] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Traceback (most recent call last): [ 740.646409] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 740.646409] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] self.driver.spawn(context, instance, image_meta, [ 740.646409] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 740.646409] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] self._vmops.spawn(context, instance, image_meta, injected_files, [ 740.646409] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 740.646409] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] vm_ref = self.build_virtual_machine(instance, [ 740.646409] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 740.646409] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] vif_infos = vmwarevif.get_vif_info(self._session, [ 740.646409] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 740.646826] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] for vif in network_info: [ 740.646826] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 740.646826] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] return self._sync_wrapper(fn, *args, **kwargs) [ 740.646826] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 740.646826] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] self.wait() [ 740.646826] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 740.646826] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] self[:] = self._gt.wait() [ 740.646826] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 740.646826] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] return self._exit_event.wait() [ 740.646826] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 740.646826] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] current.throw(*self._exc) [ 740.646826] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 740.646826] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] result = function(*args, **kwargs) [ 740.647258] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 740.647258] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] return func(*args, **kwargs) [ 740.647258] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 740.647258] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] raise e [ 740.647258] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 740.647258] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] nwinfo = self.network_api.allocate_for_instance( [ 740.647258] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 740.647258] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] created_port_ids = self._update_ports_for_instance( [ 740.647258] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 740.647258] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] with excutils.save_and_reraise_exception(): [ 740.647258] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 740.647258] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] self.force_reraise() [ 740.647258] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 740.647690] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] raise self.value [ 740.647690] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 740.647690] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] updated_port = self._update_port( [ 740.647690] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 740.647690] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] _ensure_no_port_binding_failure(port) [ 740.647690] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 740.647690] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] raise exception.PortBindingFailed(port_id=port['id']) [ 740.647690] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] nova.exception.PortBindingFailed: Binding failed for port 84318521-767d-49b3-81f5-02cbb8556c81, please check neutron logs for more information. [ 740.647690] env[62965]: ERROR nova.compute.manager [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] [ 740.647690] env[62965]: DEBUG nova.compute.utils [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Binding failed for port 84318521-767d-49b3-81f5-02cbb8556c81, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 740.648299] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.846s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.651062] env[62965]: DEBUG nova.compute.manager [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Build of instance 83c4b731-780a-4f3c-bbdd-b624a52365ef was re-scheduled: Binding failed for port 84318521-767d-49b3-81f5-02cbb8556c81, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 740.651473] env[62965]: DEBUG nova.compute.manager [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 740.651687] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "refresh_cache-83c4b731-780a-4f3c-bbdd-b624a52365ef" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.651831] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquired lock "refresh_cache-83c4b731-780a-4f3c-bbdd-b624a52365ef" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.651987] env[62965]: DEBUG nova.network.neutron [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 740.935316] env[62965]: INFO nova.compute.manager [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] [instance: b2f38e92-37a1-4e0b-9292-b20ea43a2038] Took 1.02 seconds to deallocate network for instance. [ 741.065330] env[62965]: DEBUG nova.compute.manager [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 741.180496] env[62965]: DEBUG nova.network.neutron [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 741.329641] env[62965]: DEBUG nova.network.neutron [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.491245] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e722bda-4fa2-4ca3-b50f-1d6b6f410ab2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.499988] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a6969a4-5454-46e8-947c-aae92671c50a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.531918] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d4391c3-7c82-4717-b291-958d29279fc9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.539130] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee858334-9320-4e0b-b926-8d0715c7d1aa {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.553028] env[62965]: DEBUG nova.compute.provider_tree [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 741.581921] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.832209] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Releasing lock "refresh_cache-83c4b731-780a-4f3c-bbdd-b624a52365ef" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.832455] env[62965]: DEBUG nova.compute.manager [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 741.832638] env[62965]: DEBUG nova.compute.manager [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 741.832800] env[62965]: DEBUG nova.network.neutron [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 741.852996] env[62965]: DEBUG nova.network.neutron [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 741.969381] env[62965]: INFO nova.scheduler.client.report [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] Deleted allocations for instance b2f38e92-37a1-4e0b-9292-b20ea43a2038 [ 742.061044] env[62965]: DEBUG nova.scheduler.client.report [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 742.355849] env[62965]: DEBUG nova.network.neutron [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.479708] env[62965]: DEBUG oslo_concurrency.lockutils [None req-61a1f88a-6d62-493a-bf03-042fd215ccaf tempest-ServersTestFqdnHostnames-1809024807 tempest-ServersTestFqdnHostnames-1809024807-project-member] Lock "b2f38e92-37a1-4e0b-9292-b20ea43a2038" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 155.617s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.568805] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.920s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.569457] env[62965]: ERROR nova.compute.manager [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 57373793-dcc9-4a89-8e78-d8a737cacada, please check neutron logs for more information. [ 742.569457] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Traceback (most recent call last): [ 742.569457] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 742.569457] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] self.driver.spawn(context, instance, image_meta, [ 742.569457] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 742.569457] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 742.569457] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 742.569457] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] vm_ref = self.build_virtual_machine(instance, [ 742.569457] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 742.569457] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] vif_infos = vmwarevif.get_vif_info(self._session, [ 742.569457] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 742.570042] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] for vif in network_info: [ 742.570042] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 742.570042] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] return self._sync_wrapper(fn, *args, **kwargs) [ 742.570042] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 742.570042] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] self.wait() [ 742.570042] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 742.570042] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] self[:] = self._gt.wait() [ 742.570042] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 742.570042] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] return self._exit_event.wait() [ 742.570042] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 742.570042] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] current.throw(*self._exc) [ 742.570042] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 742.570042] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] result = function(*args, **kwargs) [ 742.570682] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 742.570682] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] return func(*args, **kwargs) [ 742.570682] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 742.570682] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] raise e [ 742.570682] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 742.570682] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] nwinfo = self.network_api.allocate_for_instance( [ 742.570682] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 742.570682] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] created_port_ids = self._update_ports_for_instance( [ 742.570682] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 742.570682] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] with excutils.save_and_reraise_exception(): [ 742.570682] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 742.570682] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] self.force_reraise() [ 742.570682] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 742.571316] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] raise self.value [ 742.571316] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 742.571316] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] updated_port = self._update_port( [ 742.571316] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 742.571316] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] _ensure_no_port_binding_failure(port) [ 742.571316] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 742.571316] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] raise exception.PortBindingFailed(port_id=port['id']) [ 742.571316] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] nova.exception.PortBindingFailed: Binding failed for port 57373793-dcc9-4a89-8e78-d8a737cacada, please check neutron logs for more information. [ 742.571316] env[62965]: ERROR nova.compute.manager [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] [ 742.571316] env[62965]: DEBUG nova.compute.utils [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Binding failed for port 57373793-dcc9-4a89-8e78-d8a737cacada, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 742.571834] env[62965]: DEBUG oslo_concurrency.lockutils [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.298s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.574299] env[62965]: DEBUG nova.compute.manager [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Build of instance f8ba32e4-6234-4dee-8bc5-b7a12105a3c4 was re-scheduled: Binding failed for port 57373793-dcc9-4a89-8e78-d8a737cacada, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 742.574713] env[62965]: DEBUG nova.compute.manager [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 742.574933] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] Acquiring lock "refresh_cache-f8ba32e4-6234-4dee-8bc5-b7a12105a3c4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.575089] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] Acquired lock "refresh_cache-f8ba32e4-6234-4dee-8bc5-b7a12105a3c4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.575247] env[62965]: DEBUG nova.network.neutron [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 742.860460] env[62965]: INFO nova.compute.manager [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 83c4b731-780a-4f3c-bbdd-b624a52365ef] Took 1.03 seconds to deallocate network for instance. [ 742.987089] env[62965]: DEBUG nova.compute.manager [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 743.106492] env[62965]: DEBUG nova.network.neutron [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 743.248254] env[62965]: DEBUG nova.network.neutron [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.433961] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fcb0c9f-529c-4613-945a-d009f20ed06b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.442209] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a4a9278-6827-4bcb-8a76-781e970329b4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.472236] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee4a1a03-e477-4a96-8b26-429926b1e345 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.479881] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-366f4a74-7e99-4a90-aed7-6a6eab91f470 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.496172] env[62965]: DEBUG nova.compute.provider_tree [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 743.511391] env[62965]: DEBUG oslo_concurrency.lockutils [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 743.754624] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] Releasing lock "refresh_cache-f8ba32e4-6234-4dee-8bc5-b7a12105a3c4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.755849] env[62965]: DEBUG nova.compute.manager [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 743.756140] env[62965]: DEBUG nova.compute.manager [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 743.756317] env[62965]: DEBUG nova.network.neutron [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 743.785614] env[62965]: DEBUG nova.network.neutron [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 743.890786] env[62965]: INFO nova.scheduler.client.report [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Deleted allocations for instance 83c4b731-780a-4f3c-bbdd-b624a52365ef [ 744.000929] env[62965]: DEBUG nova.scheduler.client.report [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 744.289173] env[62965]: DEBUG nova.network.neutron [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.399318] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a43d1073-3bcb-4e01-aa8c-2afabcc5af16 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "83c4b731-780a-4f3c-bbdd-b624a52365ef" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 157.044s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.508419] env[62965]: DEBUG oslo_concurrency.lockutils [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.937s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.509217] env[62965]: ERROR nova.compute.manager [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 46f93633-3617-45b5-9054-fa6bfc49bee8, please check neutron logs for more information. [ 744.509217] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Traceback (most recent call last): [ 744.509217] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 744.509217] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] self.driver.spawn(context, instance, image_meta, [ 744.509217] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 744.509217] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 744.509217] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 744.509217] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] vm_ref = self.build_virtual_machine(instance, [ 744.509217] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 744.509217] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] vif_infos = vmwarevif.get_vif_info(self._session, [ 744.509217] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 744.510912] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] for vif in network_info: [ 744.510912] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 744.510912] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] return self._sync_wrapper(fn, *args, **kwargs) [ 744.510912] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 744.510912] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] self.wait() [ 744.510912] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 744.510912] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] self[:] = self._gt.wait() [ 744.510912] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 744.510912] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] return self._exit_event.wait() [ 744.510912] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 744.510912] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] current.throw(*self._exc) [ 744.510912] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 744.510912] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] result = function(*args, **kwargs) [ 744.511485] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 744.511485] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] return func(*args, **kwargs) [ 744.511485] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 744.511485] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] raise e [ 744.511485] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 744.511485] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] nwinfo = self.network_api.allocate_for_instance( [ 744.511485] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 744.511485] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] created_port_ids = self._update_ports_for_instance( [ 744.511485] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 744.511485] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] with excutils.save_and_reraise_exception(): [ 744.511485] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 744.511485] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] self.force_reraise() [ 744.511485] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 744.511949] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] raise self.value [ 744.511949] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 744.511949] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] updated_port = self._update_port( [ 744.511949] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 744.511949] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] _ensure_no_port_binding_failure(port) [ 744.511949] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 744.511949] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] raise exception.PortBindingFailed(port_id=port['id']) [ 744.511949] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] nova.exception.PortBindingFailed: Binding failed for port 46f93633-3617-45b5-9054-fa6bfc49bee8, please check neutron logs for more information. [ 744.511949] env[62965]: ERROR nova.compute.manager [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] [ 744.511949] env[62965]: DEBUG nova.compute.utils [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Binding failed for port 46f93633-3617-45b5-9054-fa6bfc49bee8, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 744.512415] env[62965]: DEBUG oslo_concurrency.lockutils [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.641s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.514253] env[62965]: DEBUG nova.compute.manager [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Build of instance ece118fc-a39e-443a-bfd1-fc332da2452f was re-scheduled: Binding failed for port 46f93633-3617-45b5-9054-fa6bfc49bee8, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 744.514684] env[62965]: DEBUG nova.compute.manager [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 744.518232] env[62965]: DEBUG oslo_concurrency.lockutils [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Acquiring lock "refresh_cache-ece118fc-a39e-443a-bfd1-fc332da2452f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.518232] env[62965]: DEBUG oslo_concurrency.lockutils [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Acquired lock "refresh_cache-ece118fc-a39e-443a-bfd1-fc332da2452f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.518232] env[62965]: DEBUG nova.network.neutron [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 744.792167] env[62965]: INFO nova.compute.manager [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] [instance: f8ba32e4-6234-4dee-8bc5-b7a12105a3c4] Took 1.04 seconds to deallocate network for instance. [ 744.905572] env[62965]: DEBUG nova.compute.manager [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 745.039293] env[62965]: DEBUG nova.network.neutron [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 745.164262] env[62965]: DEBUG nova.network.neutron [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.390104] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c82f5847-e90f-42b8-8469-6b907c4030fc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.399854] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a02af73-8a05-4bbb-8789-36af4df46569 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.435369] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-710d0b3f-b668-4cca-964e-537454846264 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.443269] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f8395d5-e011-4ba3-8f2b-db128e9c045d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.449307] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 745.457239] env[62965]: DEBUG nova.compute.provider_tree [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 745.666971] env[62965]: DEBUG oslo_concurrency.lockutils [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Releasing lock "refresh_cache-ece118fc-a39e-443a-bfd1-fc332da2452f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.667243] env[62965]: DEBUG nova.compute.manager [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 745.668026] env[62965]: DEBUG nova.compute.manager [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 745.668026] env[62965]: DEBUG nova.network.neutron [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 745.688435] env[62965]: DEBUG nova.network.neutron [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 745.825852] env[62965]: INFO nova.scheduler.client.report [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] Deleted allocations for instance f8ba32e4-6234-4dee-8bc5-b7a12105a3c4 [ 745.960526] env[62965]: DEBUG nova.scheduler.client.report [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 746.191927] env[62965]: DEBUG nova.network.neutron [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.334628] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e0687c9f-8fba-4d30-8338-ef402360d49d tempest-ServersNegativeTestMultiTenantJSON-1247453518 tempest-ServersNegativeTestMultiTenantJSON-1247453518-project-member] Lock "f8ba32e4-6234-4dee-8bc5-b7a12105a3c4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 158.564s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.466424] env[62965]: DEBUG oslo_concurrency.lockutils [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.955s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.467182] env[62965]: ERROR nova.compute.manager [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 046d1077-d6d9-4ae5-9a33-89a4ba0a7b08, please check neutron logs for more information. [ 746.467182] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Traceback (most recent call last): [ 746.467182] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 746.467182] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] self.driver.spawn(context, instance, image_meta, [ 746.467182] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 746.467182] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 746.467182] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 746.467182] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] vm_ref = self.build_virtual_machine(instance, [ 746.467182] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 746.467182] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] vif_infos = vmwarevif.get_vif_info(self._session, [ 746.467182] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 746.467559] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] for vif in network_info: [ 746.467559] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 746.467559] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] return self._sync_wrapper(fn, *args, **kwargs) [ 746.467559] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 746.467559] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] self.wait() [ 746.467559] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 746.467559] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] self[:] = self._gt.wait() [ 746.467559] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 746.467559] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] return self._exit_event.wait() [ 746.467559] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 746.467559] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] current.throw(*self._exc) [ 746.467559] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 746.467559] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] result = function(*args, **kwargs) [ 746.467929] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 746.467929] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] return func(*args, **kwargs) [ 746.467929] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 746.467929] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] raise e [ 746.467929] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 746.467929] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] nwinfo = self.network_api.allocate_for_instance( [ 746.467929] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 746.467929] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] created_port_ids = self._update_ports_for_instance( [ 746.467929] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 746.467929] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] with excutils.save_and_reraise_exception(): [ 746.467929] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.467929] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] self.force_reraise() [ 746.467929] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.468332] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] raise self.value [ 746.468332] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 746.468332] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] updated_port = self._update_port( [ 746.468332] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.468332] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] _ensure_no_port_binding_failure(port) [ 746.468332] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.468332] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] raise exception.PortBindingFailed(port_id=port['id']) [ 746.468332] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] nova.exception.PortBindingFailed: Binding failed for port 046d1077-d6d9-4ae5-9a33-89a4ba0a7b08, please check neutron logs for more information. [ 746.468332] env[62965]: ERROR nova.compute.manager [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] [ 746.468332] env[62965]: DEBUG nova.compute.utils [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Binding failed for port 046d1077-d6d9-4ae5-9a33-89a4ba0a7b08, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 746.469316] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.912s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.471985] env[62965]: DEBUG nova.compute.manager [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Build of instance 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb was re-scheduled: Binding failed for port 046d1077-d6d9-4ae5-9a33-89a4ba0a7b08, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 746.472480] env[62965]: DEBUG nova.compute.manager [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 746.472773] env[62965]: DEBUG oslo_concurrency.lockutils [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] Acquiring lock "refresh_cache-1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.472773] env[62965]: DEBUG oslo_concurrency.lockutils [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] Acquired lock "refresh_cache-1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.474216] env[62965]: DEBUG nova.network.neutron [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 746.694933] env[62965]: INFO nova.compute.manager [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: ece118fc-a39e-443a-bfd1-fc332da2452f] Took 1.03 seconds to deallocate network for instance. [ 746.837531] env[62965]: DEBUG nova.compute.manager [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: afaec369-6aef-4bdc-b886-eed795644867] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 747.013382] env[62965]: DEBUG nova.network.neutron [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 747.149172] env[62965]: DEBUG nova.network.neutron [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.356649] env[62965]: DEBUG oslo_concurrency.lockutils [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.367016] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b7d9b6b-b1ba-44de-a572-58cb26dde265 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.373731] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05e6caff-5e66-46fc-855a-7e98931a245c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.406236] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c3f12ae-cdf9-45fd-bbce-d129ec3dc341 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.413972] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-accb89e4-5f35-48f9-8fde-8cce4e006fac {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.428143] env[62965]: DEBUG nova.compute.provider_tree [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 747.647838] env[62965]: DEBUG oslo_concurrency.lockutils [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "9566eb77-5198-47b3-a044-d2ed34e00888" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.648265] env[62965]: DEBUG oslo_concurrency.lockutils [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "9566eb77-5198-47b3-a044-d2ed34e00888" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.652245] env[62965]: DEBUG oslo_concurrency.lockutils [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] Releasing lock "refresh_cache-1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.652581] env[62965]: DEBUG nova.compute.manager [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 747.653294] env[62965]: DEBUG nova.compute.manager [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 747.653743] env[62965]: DEBUG nova.network.neutron [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 747.684770] env[62965]: DEBUG nova.network.neutron [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 747.746520] env[62965]: INFO nova.scheduler.client.report [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Deleted allocations for instance ece118fc-a39e-443a-bfd1-fc332da2452f [ 747.934338] env[62965]: DEBUG nova.scheduler.client.report [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 748.189765] env[62965]: DEBUG nova.network.neutron [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.262787] env[62965]: DEBUG oslo_concurrency.lockutils [None req-27a53aa9-b155-4b00-9f3a-d6c73b2e1e4f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Lock "ece118fc-a39e-443a-bfd1-fc332da2452f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 159.936s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.442357] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.973s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.443463] env[62965]: ERROR nova.compute.manager [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port dc4a4c53-a1ac-4fd7-a63e-5a7031939c89, please check neutron logs for more information. [ 748.443463] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Traceback (most recent call last): [ 748.443463] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 748.443463] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] self.driver.spawn(context, instance, image_meta, [ 748.443463] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 748.443463] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 748.443463] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 748.443463] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] vm_ref = self.build_virtual_machine(instance, [ 748.443463] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 748.443463] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] vif_infos = vmwarevif.get_vif_info(self._session, [ 748.443463] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 748.443950] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] for vif in network_info: [ 748.443950] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 748.443950] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] return self._sync_wrapper(fn, *args, **kwargs) [ 748.443950] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 748.443950] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] self.wait() [ 748.443950] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 748.443950] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] self[:] = self._gt.wait() [ 748.443950] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 748.443950] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] return self._exit_event.wait() [ 748.443950] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 748.443950] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] current.throw(*self._exc) [ 748.443950] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 748.443950] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] result = function(*args, **kwargs) [ 748.444338] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 748.444338] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] return func(*args, **kwargs) [ 748.444338] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 748.444338] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] raise e [ 748.444338] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 748.444338] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] nwinfo = self.network_api.allocate_for_instance( [ 748.444338] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 748.444338] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] created_port_ids = self._update_ports_for_instance( [ 748.444338] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 748.444338] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] with excutils.save_and_reraise_exception(): [ 748.444338] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 748.444338] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] self.force_reraise() [ 748.444338] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 748.444855] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] raise self.value [ 748.444855] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 748.444855] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] updated_port = self._update_port( [ 748.444855] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 748.444855] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] _ensure_no_port_binding_failure(port) [ 748.444855] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 748.444855] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] raise exception.PortBindingFailed(port_id=port['id']) [ 748.444855] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] nova.exception.PortBindingFailed: Binding failed for port dc4a4c53-a1ac-4fd7-a63e-5a7031939c89, please check neutron logs for more information. [ 748.444855] env[62965]: ERROR nova.compute.manager [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] [ 748.444855] env[62965]: DEBUG nova.compute.utils [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Binding failed for port dc4a4c53-a1ac-4fd7-a63e-5a7031939c89, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 748.445228] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.221s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.450495] env[62965]: INFO nova.compute.claims [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 748.453729] env[62965]: DEBUG nova.compute.manager [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Build of instance 99db6bef-6501-4ab3-a329-0877ca303f1b was re-scheduled: Binding failed for port dc4a4c53-a1ac-4fd7-a63e-5a7031939c89, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 748.457068] env[62965]: DEBUG nova.compute.manager [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 748.457068] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] Acquiring lock "refresh_cache-99db6bef-6501-4ab3-a329-0877ca303f1b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.457068] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] Acquired lock "refresh_cache-99db6bef-6501-4ab3-a329-0877ca303f1b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.457068] env[62965]: DEBUG nova.network.neutron [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 748.692147] env[62965]: INFO nova.compute.manager [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] [instance: 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb] Took 1.04 seconds to deallocate network for instance. [ 748.767939] env[62965]: DEBUG nova.compute.manager [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 748.770975] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] Acquiring lock "2b6efa32-fc5b-402c-a551-b6e5c01eca2c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.770975] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] Lock "2b6efa32-fc5b-402c-a551-b6e5c01eca2c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.981522] env[62965]: DEBUG nova.network.neutron [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 749.111984] env[62965]: DEBUG nova.network.neutron [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.302512] env[62965]: DEBUG oslo_concurrency.lockutils [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.588979] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 749.588979] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 749.616565] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] Releasing lock "refresh_cache-99db6bef-6501-4ab3-a329-0877ca303f1b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.616565] env[62965]: DEBUG nova.compute.manager [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 749.616565] env[62965]: DEBUG nova.compute.manager [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 749.616741] env[62965]: DEBUG nova.network.neutron [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 749.634255] env[62965]: DEBUG nova.network.neutron [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 749.734457] env[62965]: INFO nova.scheduler.client.report [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] Deleted allocations for instance 1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb [ 749.829188] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3892267c-ede0-426c-a2b9-05d45ef52133 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.837994] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc786dc6-a34d-4bd5-983a-79dc6d62c2f1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.867588] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c477148-0410-4e22-a43a-dd399e8b8f45 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.875233] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc46dcf8-272a-4bd6-93f8-b89ab1cfab57 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.890369] env[62965]: DEBUG nova.compute.provider_tree [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 750.093505] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 750.093710] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Starting heal instance info cache {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10257}} [ 750.093831] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Rebuilding the list of instances to heal {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10261}} [ 750.136971] env[62965]: DEBUG nova.network.neutron [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.245704] env[62965]: DEBUG oslo_concurrency.lockutils [None req-65b83c48-b8b2-4d4f-abcd-f0cf6a723184 tempest-ServerActionsTestOtherB-1134441931 tempest-ServerActionsTestOtherB-1134441931-project-member] Lock "1a55a4d1-59ea-4dee-b0bd-de5d04e3ecdb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 156.027s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.395605] env[62965]: DEBUG nova.scheduler.client.report [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 750.597701] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Skipping network cache update for instance because it is Building. {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10270}} [ 750.597863] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Skipping network cache update for instance because it is Building. {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10270}} [ 750.597990] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Didn't find any instances for network info cache update. {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10343}} [ 750.598216] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 750.598364] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 750.598501] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 750.598661] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 750.600053] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 750.600053] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 750.600053] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62965) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10876}} [ 750.600053] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 750.640643] env[62965]: INFO nova.compute.manager [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] [instance: 99db6bef-6501-4ab3-a329-0877ca303f1b] Took 1.02 seconds to deallocate network for instance. [ 750.748520] env[62965]: DEBUG nova.compute.manager [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 750.900062] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.454s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.900062] env[62965]: DEBUG nova.compute.manager [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 750.902221] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.453s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.102639] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.114868] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Acquiring lock "64cf898c-2c97-4daa-80c8-0a1c252cbcf2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.114868] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Lock "64cf898c-2c97-4daa-80c8-0a1c252cbcf2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.271724] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.410108] env[62965]: DEBUG nova.compute.utils [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 751.411568] env[62965]: DEBUG nova.compute.manager [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 751.411731] env[62965]: DEBUG nova.network.neutron [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 751.489821] env[62965]: DEBUG nova.policy [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7455aa474b35492eaa9cf8c25023ff5a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e7955d27668a4872993b7e32bd445ac9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 751.679053] env[62965]: INFO nova.scheduler.client.report [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] Deleted allocations for instance 99db6bef-6501-4ab3-a329-0877ca303f1b [ 751.754448] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12b51107-9f41-4076-b56e-cb6f5fb30017 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.762538] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80e26988-4034-4cbf-b0e4-ac3ea049242e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.795623] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa9cdc37-ac6e-4ee1-8d29-63d79564ac2f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.812243] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad002924-17a9-45d5-93ae-f4476b1b723b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.823869] env[62965]: DEBUG nova.compute.provider_tree [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 751.916950] env[62965]: DEBUG nova.compute.manager [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 752.187424] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c246dded-7b57-4b15-b836-40ce6169b748 tempest-ServersTestBootFromVolume-1228806318 tempest-ServersTestBootFromVolume-1228806318-project-member] Lock "99db6bef-6501-4ab3-a329-0877ca303f1b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 154.738s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.220676] env[62965]: DEBUG nova.network.neutron [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Successfully created port: 171591e5-b76f-4852-b993-0c69d8392a0c {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 752.327420] env[62965]: DEBUG nova.scheduler.client.report [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 752.692394] env[62965]: DEBUG nova.compute.manager [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 752.837921] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.931s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.837921] env[62965]: ERROR nova.compute.manager [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7af92ac6-9925-4abe-a323-98c576e48b87, please check neutron logs for more information. [ 752.837921] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Traceback (most recent call last): [ 752.837921] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 752.837921] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] self.driver.spawn(context, instance, image_meta, [ 752.837921] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 752.837921] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] self._vmops.spawn(context, instance, image_meta, injected_files, [ 752.837921] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 752.837921] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] vm_ref = self.build_virtual_machine(instance, [ 752.838849] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 752.838849] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] vif_infos = vmwarevif.get_vif_info(self._session, [ 752.838849] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 752.838849] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] for vif in network_info: [ 752.838849] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 752.838849] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] return self._sync_wrapper(fn, *args, **kwargs) [ 752.838849] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 752.838849] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] self.wait() [ 752.838849] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 752.838849] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] self[:] = self._gt.wait() [ 752.838849] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 752.838849] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] return self._exit_event.wait() [ 752.838849] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 752.839300] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] result = hub.switch() [ 752.839300] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 752.839300] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] return self.greenlet.switch() [ 752.839300] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 752.839300] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] result = function(*args, **kwargs) [ 752.839300] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 752.839300] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] return func(*args, **kwargs) [ 752.839300] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 752.839300] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] raise e [ 752.839300] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 752.839300] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] nwinfo = self.network_api.allocate_for_instance( [ 752.839300] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 752.839300] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] created_port_ids = self._update_ports_for_instance( [ 752.839775] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 752.839775] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] with excutils.save_and_reraise_exception(): [ 752.839775] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 752.839775] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] self.force_reraise() [ 752.839775] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 752.839775] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] raise self.value [ 752.839775] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 752.839775] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] updated_port = self._update_port( [ 752.839775] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 752.839775] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] _ensure_no_port_binding_failure(port) [ 752.839775] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 752.839775] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] raise exception.PortBindingFailed(port_id=port['id']) [ 752.840220] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] nova.exception.PortBindingFailed: Binding failed for port 7af92ac6-9925-4abe-a323-98c576e48b87, please check neutron logs for more information. [ 752.840220] env[62965]: ERROR nova.compute.manager [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] [ 752.840220] env[62965]: DEBUG nova.compute.utils [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Binding failed for port 7af92ac6-9925-4abe-a323-98c576e48b87, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 752.841717] env[62965]: DEBUG nova.compute.manager [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Build of instance 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add was re-scheduled: Binding failed for port 7af92ac6-9925-4abe-a323-98c576e48b87, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 752.841717] env[62965]: DEBUG nova.compute.manager [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 752.842662] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] Acquiring lock "refresh_cache-8342b9df-ebfd-403b-a0b8-1fdd6ebe2add" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.842662] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] Acquired lock "refresh_cache-8342b9df-ebfd-403b-a0b8-1fdd6ebe2add" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.842662] env[62965]: DEBUG nova.network.neutron [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 752.843645] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.337s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.845790] env[62965]: INFO nova.compute.claims [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 752.927382] env[62965]: DEBUG nova.compute.manager [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 752.963616] env[62965]: DEBUG nova.virt.hardware [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 752.964635] env[62965]: DEBUG nova.virt.hardware [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 752.964635] env[62965]: DEBUG nova.virt.hardware [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 752.964635] env[62965]: DEBUG nova.virt.hardware [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 752.964635] env[62965]: DEBUG nova.virt.hardware [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 752.964635] env[62965]: DEBUG nova.virt.hardware [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 752.965166] env[62965]: DEBUG nova.virt.hardware [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 752.965735] env[62965]: DEBUG nova.virt.hardware [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 752.965984] env[62965]: DEBUG nova.virt.hardware [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 752.966267] env[62965]: DEBUG nova.virt.hardware [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 752.966525] env[62965]: DEBUG nova.virt.hardware [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 752.968374] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5473894b-8d27-4e57-acfe-241e13baccab {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.979778] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60e1afd1-b289-4e04-9cf5-3e532fa22a26 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.218921] env[62965]: DEBUG oslo_concurrency.lockutils [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.372765] env[62965]: DEBUG nova.network.neutron [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 753.520251] env[62965]: DEBUG nova.compute.manager [req-b80c13b6-cca8-42a2-8553-01bddc3f2c37 req-a60a0914-4916-4870-b546-6ae1c8af05bd service nova] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Received event network-changed-171591e5-b76f-4852-b993-0c69d8392a0c {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 753.520544] env[62965]: DEBUG nova.compute.manager [req-b80c13b6-cca8-42a2-8553-01bddc3f2c37 req-a60a0914-4916-4870-b546-6ae1c8af05bd service nova] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Refreshing instance network info cache due to event network-changed-171591e5-b76f-4852-b993-0c69d8392a0c. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 753.520646] env[62965]: DEBUG oslo_concurrency.lockutils [req-b80c13b6-cca8-42a2-8553-01bddc3f2c37 req-a60a0914-4916-4870-b546-6ae1c8af05bd service nova] Acquiring lock "refresh_cache-cd978d56-ba80-44a9-9a02-3a1254066571" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.520803] env[62965]: DEBUG oslo_concurrency.lockutils [req-b80c13b6-cca8-42a2-8553-01bddc3f2c37 req-a60a0914-4916-4870-b546-6ae1c8af05bd service nova] Acquired lock "refresh_cache-cd978d56-ba80-44a9-9a02-3a1254066571" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.520971] env[62965]: DEBUG nova.network.neutron [req-b80c13b6-cca8-42a2-8553-01bddc3f2c37 req-a60a0914-4916-4870-b546-6ae1c8af05bd service nova] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Refreshing network info cache for port 171591e5-b76f-4852-b993-0c69d8392a0c {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 753.529331] env[62965]: DEBUG nova.network.neutron [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.726936] env[62965]: ERROR nova.compute.manager [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 171591e5-b76f-4852-b993-0c69d8392a0c, please check neutron logs for more information. [ 753.726936] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 753.726936] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 753.726936] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 753.726936] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 753.726936] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 753.726936] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 753.726936] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 753.726936] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.726936] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 753.726936] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.726936] env[62965]: ERROR nova.compute.manager raise self.value [ 753.726936] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 753.726936] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 753.726936] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.726936] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 753.727563] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.727563] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 753.727563] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 171591e5-b76f-4852-b993-0c69d8392a0c, please check neutron logs for more information. [ 753.727563] env[62965]: ERROR nova.compute.manager [ 753.727563] env[62965]: Traceback (most recent call last): [ 753.727563] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 753.727563] env[62965]: listener.cb(fileno) [ 753.727563] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 753.727563] env[62965]: result = function(*args, **kwargs) [ 753.727563] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 753.727563] env[62965]: return func(*args, **kwargs) [ 753.727563] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 753.727563] env[62965]: raise e [ 753.727563] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 753.727563] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 753.727563] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 753.727563] env[62965]: created_port_ids = self._update_ports_for_instance( [ 753.727563] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 753.727563] env[62965]: with excutils.save_and_reraise_exception(): [ 753.727563] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.727563] env[62965]: self.force_reraise() [ 753.727563] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.727563] env[62965]: raise self.value [ 753.727563] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 753.727563] env[62965]: updated_port = self._update_port( [ 753.727563] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.727563] env[62965]: _ensure_no_port_binding_failure(port) [ 753.727563] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.727563] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 753.728636] env[62965]: nova.exception.PortBindingFailed: Binding failed for port 171591e5-b76f-4852-b993-0c69d8392a0c, please check neutron logs for more information. [ 753.728636] env[62965]: Removing descriptor: 17 [ 753.728636] env[62965]: ERROR nova.compute.manager [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 171591e5-b76f-4852-b993-0c69d8392a0c, please check neutron logs for more information. [ 753.728636] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Traceback (most recent call last): [ 753.728636] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 753.728636] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] yield resources [ 753.728636] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 753.728636] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] self.driver.spawn(context, instance, image_meta, [ 753.728636] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 753.728636] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] self._vmops.spawn(context, instance, image_meta, injected_files, [ 753.728636] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 753.728636] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] vm_ref = self.build_virtual_machine(instance, [ 753.729130] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 753.729130] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] vif_infos = vmwarevif.get_vif_info(self._session, [ 753.729130] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 753.729130] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] for vif in network_info: [ 753.729130] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 753.729130] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] return self._sync_wrapper(fn, *args, **kwargs) [ 753.729130] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 753.729130] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] self.wait() [ 753.729130] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 753.729130] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] self[:] = self._gt.wait() [ 753.729130] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 753.729130] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] return self._exit_event.wait() [ 753.729130] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 753.729606] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] result = hub.switch() [ 753.729606] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 753.729606] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] return self.greenlet.switch() [ 753.729606] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 753.729606] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] result = function(*args, **kwargs) [ 753.729606] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 753.729606] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] return func(*args, **kwargs) [ 753.729606] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 753.729606] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] raise e [ 753.729606] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 753.729606] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] nwinfo = self.network_api.allocate_for_instance( [ 753.729606] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 753.729606] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] created_port_ids = self._update_ports_for_instance( [ 753.730500] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 753.730500] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] with excutils.save_and_reraise_exception(): [ 753.730500] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 753.730500] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] self.force_reraise() [ 753.730500] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 753.730500] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] raise self.value [ 753.730500] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 753.730500] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] updated_port = self._update_port( [ 753.730500] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 753.730500] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] _ensure_no_port_binding_failure(port) [ 753.730500] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 753.730500] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] raise exception.PortBindingFailed(port_id=port['id']) [ 753.731210] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] nova.exception.PortBindingFailed: Binding failed for port 171591e5-b76f-4852-b993-0c69d8392a0c, please check neutron logs for more information. [ 753.731210] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] [ 753.731210] env[62965]: INFO nova.compute.manager [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Terminating instance [ 754.035659] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] Releasing lock "refresh_cache-8342b9df-ebfd-403b-a0b8-1fdd6ebe2add" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.035659] env[62965]: DEBUG nova.compute.manager [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 754.035659] env[62965]: DEBUG nova.compute.manager [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 754.035659] env[62965]: DEBUG nova.network.neutron [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 754.044016] env[62965]: DEBUG nova.network.neutron [req-b80c13b6-cca8-42a2-8553-01bddc3f2c37 req-a60a0914-4916-4870-b546-6ae1c8af05bd service nova] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 754.058497] env[62965]: DEBUG nova.network.neutron [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 754.122577] env[62965]: DEBUG nova.network.neutron [req-b80c13b6-cca8-42a2-8553-01bddc3f2c37 req-a60a0914-4916-4870-b546-6ae1c8af05bd service nova] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.235879] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] Acquiring lock "refresh_cache-cd978d56-ba80-44a9-9a02-3a1254066571" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.253527] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80f3f12e-c324-4ffe-8316-e49640ad96e6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.261212] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24b2bce9-06ef-495d-965d-1f4fc7bf572a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.293561] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32c861f0-30b4-4cfc-b432-b0fb3ac9e0e1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.301710] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a48fdbf-2fe1-482c-8c6c-c802a0dc69eb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.316831] env[62965]: DEBUG nova.compute.provider_tree [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 754.561140] env[62965]: DEBUG nova.network.neutron [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.583415] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Acquiring lock "d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.583646] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Lock "d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.629262] env[62965]: DEBUG oslo_concurrency.lockutils [req-b80c13b6-cca8-42a2-8553-01bddc3f2c37 req-a60a0914-4916-4870-b546-6ae1c8af05bd service nova] Releasing lock "refresh_cache-cd978d56-ba80-44a9-9a02-3a1254066571" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.629813] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] Acquired lock "refresh_cache-cd978d56-ba80-44a9-9a02-3a1254066571" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.630119] env[62965]: DEBUG nova.network.neutron [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 754.820731] env[62965]: DEBUG nova.scheduler.client.report [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 755.064052] env[62965]: INFO nova.compute.manager [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] [instance: 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add] Took 1.03 seconds to deallocate network for instance. [ 755.144991] env[62965]: DEBUG nova.network.neutron [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 755.204174] env[62965]: DEBUG nova.network.neutron [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.325827] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.482s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.326353] env[62965]: DEBUG nova.compute.manager [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 755.328865] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.747s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.330548] env[62965]: INFO nova.compute.claims [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 755.544494] env[62965]: DEBUG nova.compute.manager [req-8ec7caec-1791-47cf-b2dd-96d885c6f6a3 req-5a095f4b-1017-4707-b5d1-6e397ec5e64c service nova] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Received event network-vif-deleted-171591e5-b76f-4852-b993-0c69d8392a0c {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 755.707652] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] Releasing lock "refresh_cache-cd978d56-ba80-44a9-9a02-3a1254066571" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 755.708114] env[62965]: DEBUG nova.compute.manager [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 755.708305] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 755.708588] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-edf978e6-8e4a-4573-9e2c-12da6fef388d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.717557] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d3f8284-c9a0-4d03-b085-3979739f7e42 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.739588] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance cd978d56-ba80-44a9-9a02-3a1254066571 could not be found. [ 755.739792] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 755.739967] env[62965]: INFO nova.compute.manager [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Took 0.03 seconds to destroy the instance on the hypervisor. [ 755.740216] env[62965]: DEBUG oslo.service.loopingcall [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 755.740418] env[62965]: DEBUG nova.compute.manager [-] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 755.740511] env[62965]: DEBUG nova.network.neutron [-] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 755.754346] env[62965]: DEBUG nova.network.neutron [-] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 755.834336] env[62965]: DEBUG nova.compute.utils [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 755.839189] env[62965]: DEBUG nova.compute.manager [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 755.839354] env[62965]: DEBUG nova.network.neutron [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 755.887801] env[62965]: DEBUG nova.policy [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f50a3f6cb8574392890da5144157d996', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fb1907c8137f4bc1ae2b8337d3606b53', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 756.110252] env[62965]: INFO nova.scheduler.client.report [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] Deleted allocations for instance 8342b9df-ebfd-403b-a0b8-1fdd6ebe2add [ 756.220488] env[62965]: DEBUG nova.network.neutron [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Successfully created port: 7c541fe7-1952-4a2e-8a49-bc41ca6ef389 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 756.256638] env[62965]: DEBUG nova.network.neutron [-] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.338192] env[62965]: DEBUG nova.compute.manager [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 756.620406] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3802ebbd-8df3-45d4-ae46-ece4ec547998 tempest-ServerRescueTestJSON-100932118 tempest-ServerRescueTestJSON-100932118-project-member] Lock "8342b9df-ebfd-403b-a0b8-1fdd6ebe2add" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 150.654s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.634761] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-581d0927-7cec-46f9-babc-5cebabbf6ee6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.646771] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3480bd18-10d7-451a-bf17-97386db1cbd0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.678046] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9508caf1-1480-467e-aea3-f4753d4abd9b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.685441] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-522cd077-a707-4ce0-9402-2b8f97c9eabe {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.702493] env[62965]: DEBUG nova.compute.provider_tree [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 756.760234] env[62965]: INFO nova.compute.manager [-] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Took 1.02 seconds to deallocate network for instance. [ 756.762693] env[62965]: DEBUG nova.compute.claims [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 756.762903] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.124127] env[62965]: DEBUG nova.compute.manager [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 757.206103] env[62965]: DEBUG nova.scheduler.client.report [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 757.303418] env[62965]: ERROR nova.compute.manager [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7c541fe7-1952-4a2e-8a49-bc41ca6ef389, please check neutron logs for more information. [ 757.303418] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 757.303418] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 757.303418] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 757.303418] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 757.303418] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 757.303418] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 757.303418] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 757.303418] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 757.303418] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 757.303418] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 757.303418] env[62965]: ERROR nova.compute.manager raise self.value [ 757.303418] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 757.303418] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 757.303418] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 757.303418] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 757.305305] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 757.305305] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 757.305305] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7c541fe7-1952-4a2e-8a49-bc41ca6ef389, please check neutron logs for more information. [ 757.305305] env[62965]: ERROR nova.compute.manager [ 757.305305] env[62965]: Traceback (most recent call last): [ 757.305305] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 757.305305] env[62965]: listener.cb(fileno) [ 757.305305] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 757.305305] env[62965]: result = function(*args, **kwargs) [ 757.305305] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 757.305305] env[62965]: return func(*args, **kwargs) [ 757.305305] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 757.305305] env[62965]: raise e [ 757.305305] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 757.305305] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 757.305305] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 757.305305] env[62965]: created_port_ids = self._update_ports_for_instance( [ 757.305305] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 757.305305] env[62965]: with excutils.save_and_reraise_exception(): [ 757.305305] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 757.305305] env[62965]: self.force_reraise() [ 757.305305] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 757.305305] env[62965]: raise self.value [ 757.305305] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 757.305305] env[62965]: updated_port = self._update_port( [ 757.305305] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 757.305305] env[62965]: _ensure_no_port_binding_failure(port) [ 757.305305] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 757.305305] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 757.306275] env[62965]: nova.exception.PortBindingFailed: Binding failed for port 7c541fe7-1952-4a2e-8a49-bc41ca6ef389, please check neutron logs for more information. [ 757.306275] env[62965]: Removing descriptor: 17 [ 757.346229] env[62965]: DEBUG nova.compute.manager [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 757.380370] env[62965]: DEBUG nova.virt.hardware [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 757.380618] env[62965]: DEBUG nova.virt.hardware [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 757.380770] env[62965]: DEBUG nova.virt.hardware [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 757.380948] env[62965]: DEBUG nova.virt.hardware [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 757.381151] env[62965]: DEBUG nova.virt.hardware [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 757.381272] env[62965]: DEBUG nova.virt.hardware [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 757.381482] env[62965]: DEBUG nova.virt.hardware [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 757.381633] env[62965]: DEBUG nova.virt.hardware [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 757.381800] env[62965]: DEBUG nova.virt.hardware [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 757.381964] env[62965]: DEBUG nova.virt.hardware [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 757.382157] env[62965]: DEBUG nova.virt.hardware [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 757.383054] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7e7f238-1957-47be-86ac-ccf8cb84b716 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.394359] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8741b5e3-7339-4bd6-abe6-02d30a5265d2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.409168] env[62965]: ERROR nova.compute.manager [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7c541fe7-1952-4a2e-8a49-bc41ca6ef389, please check neutron logs for more information. [ 757.409168] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Traceback (most recent call last): [ 757.409168] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 757.409168] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] yield resources [ 757.409168] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 757.409168] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] self.driver.spawn(context, instance, image_meta, [ 757.409168] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 757.409168] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 757.409168] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 757.409168] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] vm_ref = self.build_virtual_machine(instance, [ 757.409168] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 757.409710] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] vif_infos = vmwarevif.get_vif_info(self._session, [ 757.409710] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 757.409710] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] for vif in network_info: [ 757.409710] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 757.409710] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] return self._sync_wrapper(fn, *args, **kwargs) [ 757.409710] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 757.409710] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] self.wait() [ 757.409710] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 757.409710] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] self[:] = self._gt.wait() [ 757.409710] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 757.409710] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] return self._exit_event.wait() [ 757.409710] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 757.409710] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] current.throw(*self._exc) [ 757.410156] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 757.410156] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] result = function(*args, **kwargs) [ 757.410156] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 757.410156] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] return func(*args, **kwargs) [ 757.410156] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 757.410156] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] raise e [ 757.410156] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 757.410156] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] nwinfo = self.network_api.allocate_for_instance( [ 757.410156] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 757.410156] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] created_port_ids = self._update_ports_for_instance( [ 757.410156] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 757.410156] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] with excutils.save_and_reraise_exception(): [ 757.410156] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 757.410725] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] self.force_reraise() [ 757.410725] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 757.410725] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] raise self.value [ 757.410725] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 757.410725] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] updated_port = self._update_port( [ 757.410725] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 757.410725] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] _ensure_no_port_binding_failure(port) [ 757.410725] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 757.410725] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] raise exception.PortBindingFailed(port_id=port['id']) [ 757.410725] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] nova.exception.PortBindingFailed: Binding failed for port 7c541fe7-1952-4a2e-8a49-bc41ca6ef389, please check neutron logs for more information. [ 757.410725] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] [ 757.410725] env[62965]: INFO nova.compute.manager [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Terminating instance [ 757.648808] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.705143] env[62965]: DEBUG nova.compute.manager [req-21fc39d9-360c-4cf7-8b2f-9ec73dd37abc req-1b6084c0-82cb-44ae-af53-2afb382c2e55 service nova] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Received event network-changed-7c541fe7-1952-4a2e-8a49-bc41ca6ef389 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 757.705143] env[62965]: DEBUG nova.compute.manager [req-21fc39d9-360c-4cf7-8b2f-9ec73dd37abc req-1b6084c0-82cb-44ae-af53-2afb382c2e55 service nova] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Refreshing instance network info cache due to event network-changed-7c541fe7-1952-4a2e-8a49-bc41ca6ef389. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 757.705143] env[62965]: DEBUG oslo_concurrency.lockutils [req-21fc39d9-360c-4cf7-8b2f-9ec73dd37abc req-1b6084c0-82cb-44ae-af53-2afb382c2e55 service nova] Acquiring lock "refresh_cache-f9a2b957-a312-4c9b-bb0d-0b1dad6416cd" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.705143] env[62965]: DEBUG oslo_concurrency.lockutils [req-21fc39d9-360c-4cf7-8b2f-9ec73dd37abc req-1b6084c0-82cb-44ae-af53-2afb382c2e55 service nova] Acquired lock "refresh_cache-f9a2b957-a312-4c9b-bb0d-0b1dad6416cd" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.705143] env[62965]: DEBUG nova.network.neutron [req-21fc39d9-360c-4cf7-8b2f-9ec73dd37abc req-1b6084c0-82cb-44ae-af53-2afb382c2e55 service nova] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Refreshing network info cache for port 7c541fe7-1952-4a2e-8a49-bc41ca6ef389 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 757.712019] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.380s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.712019] env[62965]: DEBUG nova.compute.manager [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 757.713015] env[62965]: DEBUG oslo_concurrency.lockutils [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.202s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.717105] env[62965]: INFO nova.compute.claims [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 757.916755] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Acquiring lock "refresh_cache-f9a2b957-a312-4c9b-bb0d-0b1dad6416cd" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.214543] env[62965]: DEBUG nova.compute.utils [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 758.215930] env[62965]: DEBUG nova.compute.manager [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 758.216114] env[62965]: DEBUG nova.network.neutron [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 758.224386] env[62965]: DEBUG nova.network.neutron [req-21fc39d9-360c-4cf7-8b2f-9ec73dd37abc req-1b6084c0-82cb-44ae-af53-2afb382c2e55 service nova] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 758.281569] env[62965]: DEBUG nova.policy [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ce0e7f3429454fd18d54cbb964dfb561', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e1c6e8403409438f9ff10235cd2533bc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 758.351793] env[62965]: DEBUG nova.network.neutron [req-21fc39d9-360c-4cf7-8b2f-9ec73dd37abc req-1b6084c0-82cb-44ae-af53-2afb382c2e55 service nova] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.680728] env[62965]: DEBUG nova.network.neutron [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Successfully created port: c535343e-d317-4d03-ad15-9609c7687cac {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 758.719080] env[62965]: DEBUG nova.compute.manager [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 758.854798] env[62965]: DEBUG oslo_concurrency.lockutils [req-21fc39d9-360c-4cf7-8b2f-9ec73dd37abc req-1b6084c0-82cb-44ae-af53-2afb382c2e55 service nova] Releasing lock "refresh_cache-f9a2b957-a312-4c9b-bb0d-0b1dad6416cd" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.854798] env[62965]: DEBUG nova.compute.manager [req-21fc39d9-360c-4cf7-8b2f-9ec73dd37abc req-1b6084c0-82cb-44ae-af53-2afb382c2e55 service nova] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Received event network-vif-deleted-7c541fe7-1952-4a2e-8a49-bc41ca6ef389 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 758.854908] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Acquired lock "refresh_cache-f9a2b957-a312-4c9b-bb0d-0b1dad6416cd" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.854980] env[62965]: DEBUG nova.network.neutron [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 759.046624] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-235c01f5-c489-4319-af94-b4c744fa0935 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.054278] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be92194b-0dc4-43c7-8dd5-4d46eb16e1bf {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.083924] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a4df4d6-543b-4ee1-8a3b-5529f02c652f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.091051] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c32d46a-22e8-4e3d-a1f7-3857dc4fcb81 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.104276] env[62965]: DEBUG nova.compute.provider_tree [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 759.389125] env[62965]: DEBUG nova.network.neutron [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 759.449167] env[62965]: DEBUG nova.compute.manager [req-1aa8d5e9-1ea4-48c4-894e-331010c54177 req-709d4683-4733-4aae-9a0f-e3f2420ee23c service nova] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Received event network-changed-c535343e-d317-4d03-ad15-9609c7687cac {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 759.449167] env[62965]: DEBUG nova.compute.manager [req-1aa8d5e9-1ea4-48c4-894e-331010c54177 req-709d4683-4733-4aae-9a0f-e3f2420ee23c service nova] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Refreshing instance network info cache due to event network-changed-c535343e-d317-4d03-ad15-9609c7687cac. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 759.449167] env[62965]: DEBUG oslo_concurrency.lockutils [req-1aa8d5e9-1ea4-48c4-894e-331010c54177 req-709d4683-4733-4aae-9a0f-e3f2420ee23c service nova] Acquiring lock "refresh_cache-3a994542-8c8e-4b3c-87bc-2c40a0fc95b5" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.449308] env[62965]: DEBUG oslo_concurrency.lockutils [req-1aa8d5e9-1ea4-48c4-894e-331010c54177 req-709d4683-4733-4aae-9a0f-e3f2420ee23c service nova] Acquired lock "refresh_cache-3a994542-8c8e-4b3c-87bc-2c40a0fc95b5" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.449432] env[62965]: DEBUG nova.network.neutron [req-1aa8d5e9-1ea4-48c4-894e-331010c54177 req-709d4683-4733-4aae-9a0f-e3f2420ee23c service nova] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Refreshing network info cache for port c535343e-d317-4d03-ad15-9609c7687cac {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 759.534299] env[62965]: DEBUG nova.network.neutron [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.607216] env[62965]: DEBUG nova.scheduler.client.report [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 759.617543] env[62965]: ERROR nova.compute.manager [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c535343e-d317-4d03-ad15-9609c7687cac, please check neutron logs for more information. [ 759.617543] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 759.617543] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 759.617543] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 759.617543] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 759.617543] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 759.617543] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 759.617543] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 759.617543] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 759.617543] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 759.617543] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 759.617543] env[62965]: ERROR nova.compute.manager raise self.value [ 759.617543] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 759.617543] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 759.617543] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 759.617543] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 759.618123] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 759.618123] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 759.618123] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c535343e-d317-4d03-ad15-9609c7687cac, please check neutron logs for more information. [ 759.618123] env[62965]: ERROR nova.compute.manager [ 759.618123] env[62965]: Traceback (most recent call last): [ 759.618123] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 759.618123] env[62965]: listener.cb(fileno) [ 759.618123] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 759.618123] env[62965]: result = function(*args, **kwargs) [ 759.618123] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 759.618123] env[62965]: return func(*args, **kwargs) [ 759.618123] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 759.618123] env[62965]: raise e [ 759.618123] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 759.618123] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 759.618123] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 759.618123] env[62965]: created_port_ids = self._update_ports_for_instance( [ 759.618123] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 759.618123] env[62965]: with excutils.save_and_reraise_exception(): [ 759.618123] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 759.618123] env[62965]: self.force_reraise() [ 759.618123] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 759.618123] env[62965]: raise self.value [ 759.618123] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 759.618123] env[62965]: updated_port = self._update_port( [ 759.618123] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 759.618123] env[62965]: _ensure_no_port_binding_failure(port) [ 759.618123] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 759.618123] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 759.619117] env[62965]: nova.exception.PortBindingFailed: Binding failed for port c535343e-d317-4d03-ad15-9609c7687cac, please check neutron logs for more information. [ 759.619117] env[62965]: Removing descriptor: 19 [ 759.735049] env[62965]: DEBUG nova.compute.manager [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 759.758576] env[62965]: DEBUG nova.virt.hardware [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 759.758831] env[62965]: DEBUG nova.virt.hardware [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 759.759009] env[62965]: DEBUG nova.virt.hardware [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 759.759742] env[62965]: DEBUG nova.virt.hardware [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 759.759742] env[62965]: DEBUG nova.virt.hardware [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 759.759742] env[62965]: DEBUG nova.virt.hardware [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 759.759742] env[62965]: DEBUG nova.virt.hardware [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 759.759967] env[62965]: DEBUG nova.virt.hardware [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 759.759967] env[62965]: DEBUG nova.virt.hardware [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 759.760142] env[62965]: DEBUG nova.virt.hardware [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 759.760311] env[62965]: DEBUG nova.virt.hardware [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 759.761344] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43eaa64d-d4aa-4bd0-bd96-2d68dc958d7d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.769412] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b951a85f-f711-4908-8aee-8cd489254fb4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.783156] env[62965]: ERROR nova.compute.manager [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c535343e-d317-4d03-ad15-9609c7687cac, please check neutron logs for more information. [ 759.783156] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Traceback (most recent call last): [ 759.783156] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 759.783156] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] yield resources [ 759.783156] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 759.783156] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] self.driver.spawn(context, instance, image_meta, [ 759.783156] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 759.783156] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 759.783156] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 759.783156] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] vm_ref = self.build_virtual_machine(instance, [ 759.783156] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 759.783586] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] vif_infos = vmwarevif.get_vif_info(self._session, [ 759.783586] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 759.783586] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] for vif in network_info: [ 759.783586] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 759.783586] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] return self._sync_wrapper(fn, *args, **kwargs) [ 759.783586] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 759.783586] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] self.wait() [ 759.783586] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 759.783586] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] self[:] = self._gt.wait() [ 759.783586] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 759.783586] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] return self._exit_event.wait() [ 759.783586] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 759.783586] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] current.throw(*self._exc) [ 759.784043] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 759.784043] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] result = function(*args, **kwargs) [ 759.784043] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 759.784043] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] return func(*args, **kwargs) [ 759.784043] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 759.784043] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] raise e [ 759.784043] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 759.784043] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] nwinfo = self.network_api.allocate_for_instance( [ 759.784043] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 759.784043] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] created_port_ids = self._update_ports_for_instance( [ 759.784043] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 759.784043] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] with excutils.save_and_reraise_exception(): [ 759.784043] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 759.784501] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] self.force_reraise() [ 759.784501] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 759.784501] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] raise self.value [ 759.784501] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 759.784501] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] updated_port = self._update_port( [ 759.784501] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 759.784501] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] _ensure_no_port_binding_failure(port) [ 759.784501] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 759.784501] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] raise exception.PortBindingFailed(port_id=port['id']) [ 759.784501] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] nova.exception.PortBindingFailed: Binding failed for port c535343e-d317-4d03-ad15-9609c7687cac, please check neutron logs for more information. [ 759.784501] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] [ 759.784501] env[62965]: INFO nova.compute.manager [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Terminating instance [ 759.966335] env[62965]: DEBUG nova.network.neutron [req-1aa8d5e9-1ea4-48c4-894e-331010c54177 req-709d4683-4733-4aae-9a0f-e3f2420ee23c service nova] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 760.038057] env[62965]: DEBUG nova.network.neutron [req-1aa8d5e9-1ea4-48c4-894e-331010c54177 req-709d4683-4733-4aae-9a0f-e3f2420ee23c service nova] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.040253] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Releasing lock "refresh_cache-f9a2b957-a312-4c9b-bb0d-0b1dad6416cd" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.040483] env[62965]: DEBUG nova.compute.manager [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 760.040723] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 760.041011] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-642e7272-3de8-4b9f-aa0d-c42fa1043f53 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.049679] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab16d2d0-c448-4141-9c7a-4a9cfd96b552 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.073024] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f9a2b957-a312-4c9b-bb0d-0b1dad6416cd could not be found. [ 760.073024] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 760.073024] env[62965]: INFO nova.compute.manager [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Took 0.03 seconds to destroy the instance on the hypervisor. [ 760.073024] env[62965]: DEBUG oslo.service.loopingcall [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 760.073024] env[62965]: DEBUG nova.compute.manager [-] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 760.073024] env[62965]: DEBUG nova.network.neutron [-] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 760.087676] env[62965]: DEBUG nova.network.neutron [-] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 760.111601] env[62965]: DEBUG oslo_concurrency.lockutils [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.399s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.112097] env[62965]: DEBUG nova.compute.manager [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 760.114772] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.666s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.116202] env[62965]: INFO nova.compute.claims [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 760.287369] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquiring lock "refresh_cache-3a994542-8c8e-4b3c-87bc-2c40a0fc95b5" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.541031] env[62965]: DEBUG oslo_concurrency.lockutils [req-1aa8d5e9-1ea4-48c4-894e-331010c54177 req-709d4683-4733-4aae-9a0f-e3f2420ee23c service nova] Releasing lock "refresh_cache-3a994542-8c8e-4b3c-87bc-2c40a0fc95b5" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.541031] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquired lock "refresh_cache-3a994542-8c8e-4b3c-87bc-2c40a0fc95b5" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.541358] env[62965]: DEBUG nova.network.neutron [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 760.590855] env[62965]: DEBUG nova.network.neutron [-] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.616748] env[62965]: DEBUG nova.compute.utils [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 760.618058] env[62965]: DEBUG nova.compute.manager [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 760.618231] env[62965]: DEBUG nova.network.neutron [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: 864ce643-8620-4405-b345-9164c81fec2a] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 760.691735] env[62965]: DEBUG nova.policy [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd0d034d4165f470f8a6b28babd2d8501', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '729dd7671dd04584a35b783b99e1e3e3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 761.063869] env[62965]: DEBUG nova.network.neutron [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 761.078617] env[62965]: DEBUG nova.network.neutron [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Successfully created port: f3d10573-dcc2-4f2b-b5bd-09bddb6d75d7 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 761.095200] env[62965]: INFO nova.compute.manager [-] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Took 1.02 seconds to deallocate network for instance. [ 761.097608] env[62965]: DEBUG nova.compute.claims [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 761.098346] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.121788] env[62965]: DEBUG nova.compute.manager [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 761.170594] env[62965]: DEBUG nova.network.neutron [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.453294] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d83c3f9e-ade8-444a-ac20-c552d1805768 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.458171] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cc346e1-d5f5-497e-9135-430f0e09e64a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.489627] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c68dc282-4775-46c9-9ecd-da012cca7596 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.493105] env[62965]: DEBUG nova.compute.manager [req-16609eed-f659-444e-b7b0-52b6f04a1c71 req-cb546676-4c42-46d1-8954-5bd1c10b2b06 service nova] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Received event network-vif-deleted-c535343e-d317-4d03-ad15-9609c7687cac {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 761.499303] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2355cae-b2c5-4c07-a8d5-7aef5d029df8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.512976] env[62965]: DEBUG nova.compute.provider_tree [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 761.672490] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Releasing lock "refresh_cache-3a994542-8c8e-4b3c-87bc-2c40a0fc95b5" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.672925] env[62965]: DEBUG nova.compute.manager [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 761.673081] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 761.673361] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-45d1525e-7ae0-487b-b8a1-98c1f057469a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.682139] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64335c5f-ed6b-4996-a6b3-2540781d819e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.707892] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5 could not be found. [ 761.708131] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 761.708313] env[62965]: INFO nova.compute.manager [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Took 0.04 seconds to destroy the instance on the hypervisor. [ 761.708549] env[62965]: DEBUG oslo.service.loopingcall [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 761.709939] env[62965]: DEBUG nova.compute.manager [-] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 761.710136] env[62965]: DEBUG nova.network.neutron [-] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 761.729584] env[62965]: DEBUG nova.network.neutron [-] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 761.954909] env[62965]: ERROR nova.compute.manager [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f3d10573-dcc2-4f2b-b5bd-09bddb6d75d7, please check neutron logs for more information. [ 761.954909] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 761.954909] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 761.954909] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 761.954909] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 761.954909] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 761.954909] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 761.954909] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 761.954909] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.954909] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 761.954909] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.954909] env[62965]: ERROR nova.compute.manager raise self.value [ 761.954909] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 761.954909] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 761.954909] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.954909] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 761.955567] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.955567] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 761.955567] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f3d10573-dcc2-4f2b-b5bd-09bddb6d75d7, please check neutron logs for more information. [ 761.955567] env[62965]: ERROR nova.compute.manager [ 761.955567] env[62965]: Traceback (most recent call last): [ 761.955567] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 761.955567] env[62965]: listener.cb(fileno) [ 761.955567] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 761.955567] env[62965]: result = function(*args, **kwargs) [ 761.955567] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 761.955567] env[62965]: return func(*args, **kwargs) [ 761.955567] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 761.955567] env[62965]: raise e [ 761.955567] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 761.955567] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 761.955567] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 761.955567] env[62965]: created_port_ids = self._update_ports_for_instance( [ 761.955567] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 761.955567] env[62965]: with excutils.save_and_reraise_exception(): [ 761.955567] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.955567] env[62965]: self.force_reraise() [ 761.955567] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.955567] env[62965]: raise self.value [ 761.955567] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 761.955567] env[62965]: updated_port = self._update_port( [ 761.955567] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.955567] env[62965]: _ensure_no_port_binding_failure(port) [ 761.955567] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.955567] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 761.956907] env[62965]: nova.exception.PortBindingFailed: Binding failed for port f3d10573-dcc2-4f2b-b5bd-09bddb6d75d7, please check neutron logs for more information. [ 761.956907] env[62965]: Removing descriptor: 19 [ 762.016031] env[62965]: DEBUG nova.scheduler.client.report [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 762.135369] env[62965]: DEBUG nova.compute.manager [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 762.159408] env[62965]: DEBUG nova.virt.hardware [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 762.159641] env[62965]: DEBUG nova.virt.hardware [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 762.159795] env[62965]: DEBUG nova.virt.hardware [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 762.160010] env[62965]: DEBUG nova.virt.hardware [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 762.160166] env[62965]: DEBUG nova.virt.hardware [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 762.160312] env[62965]: DEBUG nova.virt.hardware [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 762.160511] env[62965]: DEBUG nova.virt.hardware [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 762.160664] env[62965]: DEBUG nova.virt.hardware [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 762.160827] env[62965]: DEBUG nova.virt.hardware [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 762.160983] env[62965]: DEBUG nova.virt.hardware [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 762.161184] env[62965]: DEBUG nova.virt.hardware [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 762.162320] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40297a17-1324-4f96-b3f2-bbb7a0e3c9ff {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.170179] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b56b67a7-d419-4e1e-9ffb-957cf35bedb8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.184352] env[62965]: ERROR nova.compute.manager [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f3d10573-dcc2-4f2b-b5bd-09bddb6d75d7, please check neutron logs for more information. [ 762.184352] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] Traceback (most recent call last): [ 762.184352] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 762.184352] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] yield resources [ 762.184352] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 762.184352] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] self.driver.spawn(context, instance, image_meta, [ 762.184352] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 762.184352] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 762.184352] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 762.184352] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] vm_ref = self.build_virtual_machine(instance, [ 762.184352] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 762.184786] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] vif_infos = vmwarevif.get_vif_info(self._session, [ 762.184786] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 762.184786] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] for vif in network_info: [ 762.184786] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 762.184786] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] return self._sync_wrapper(fn, *args, **kwargs) [ 762.184786] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 762.184786] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] self.wait() [ 762.184786] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 762.184786] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] self[:] = self._gt.wait() [ 762.184786] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 762.184786] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] return self._exit_event.wait() [ 762.184786] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 762.184786] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] current.throw(*self._exc) [ 762.185243] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 762.185243] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] result = function(*args, **kwargs) [ 762.185243] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 762.185243] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] return func(*args, **kwargs) [ 762.185243] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 762.185243] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] raise e [ 762.185243] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 762.185243] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] nwinfo = self.network_api.allocate_for_instance( [ 762.185243] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 762.185243] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] created_port_ids = self._update_ports_for_instance( [ 762.185243] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 762.185243] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] with excutils.save_and_reraise_exception(): [ 762.185243] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 762.185704] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] self.force_reraise() [ 762.185704] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 762.185704] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] raise self.value [ 762.185704] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 762.185704] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] updated_port = self._update_port( [ 762.185704] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 762.185704] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] _ensure_no_port_binding_failure(port) [ 762.185704] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 762.185704] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] raise exception.PortBindingFailed(port_id=port['id']) [ 762.185704] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] nova.exception.PortBindingFailed: Binding failed for port f3d10573-dcc2-4f2b-b5bd-09bddb6d75d7, please check neutron logs for more information. [ 762.185704] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] [ 762.185704] env[62965]: INFO nova.compute.manager [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Terminating instance [ 762.231656] env[62965]: DEBUG nova.network.neutron [-] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.521089] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.406s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.521585] env[62965]: DEBUG nova.compute.manager [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 762.524102] env[62965]: DEBUG oslo_concurrency.lockutils [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.168s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.525426] env[62965]: INFO nova.compute.claims [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: afaec369-6aef-4bdc-b886-eed795644867] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 762.688546] env[62965]: DEBUG oslo_concurrency.lockutils [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Acquiring lock "refresh_cache-864ce643-8620-4405-b345-9164c81fec2a" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.688913] env[62965]: DEBUG oslo_concurrency.lockutils [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Acquired lock "refresh_cache-864ce643-8620-4405-b345-9164c81fec2a" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.688964] env[62965]: DEBUG nova.network.neutron [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 762.734769] env[62965]: INFO nova.compute.manager [-] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Took 1.02 seconds to deallocate network for instance. [ 762.736608] env[62965]: DEBUG nova.compute.claims [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 762.736829] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.029483] env[62965]: DEBUG nova.compute.utils [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 763.032817] env[62965]: DEBUG nova.compute.manager [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 763.032983] env[62965]: DEBUG nova.network.neutron [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 763.094058] env[62965]: DEBUG nova.policy [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd0d034d4165f470f8a6b28babd2d8501', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '729dd7671dd04584a35b783b99e1e3e3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 763.218031] env[62965]: DEBUG nova.network.neutron [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 763.264124] env[62965]: DEBUG nova.network.neutron [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.339151] env[62965]: DEBUG nova.network.neutron [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Successfully created port: b60bf367-f412-4721-bb9f-2112fa9042de {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 763.502757] env[62965]: DEBUG nova.compute.manager [req-83c152f4-1a70-4157-9215-d7889fa13b9a req-457535cd-999b-42e5-b589-669d9f1c0057 service nova] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Received event network-changed-f3d10573-dcc2-4f2b-b5bd-09bddb6d75d7 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 763.502958] env[62965]: DEBUG nova.compute.manager [req-83c152f4-1a70-4157-9215-d7889fa13b9a req-457535cd-999b-42e5-b589-669d9f1c0057 service nova] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Refreshing instance network info cache due to event network-changed-f3d10573-dcc2-4f2b-b5bd-09bddb6d75d7. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 763.503171] env[62965]: DEBUG oslo_concurrency.lockutils [req-83c152f4-1a70-4157-9215-d7889fa13b9a req-457535cd-999b-42e5-b589-669d9f1c0057 service nova] Acquiring lock "refresh_cache-864ce643-8620-4405-b345-9164c81fec2a" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.533867] env[62965]: DEBUG nova.compute.manager [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 763.764106] env[62965]: DEBUG oslo_concurrency.lockutils [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Releasing lock "refresh_cache-864ce643-8620-4405-b345-9164c81fec2a" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.764531] env[62965]: DEBUG nova.compute.manager [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 763.764717] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 763.765028] env[62965]: DEBUG oslo_concurrency.lockutils [req-83c152f4-1a70-4157-9215-d7889fa13b9a req-457535cd-999b-42e5-b589-669d9f1c0057 service nova] Acquired lock "refresh_cache-864ce643-8620-4405-b345-9164c81fec2a" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.765201] env[62965]: DEBUG nova.network.neutron [req-83c152f4-1a70-4157-9215-d7889fa13b9a req-457535cd-999b-42e5-b589-669d9f1c0057 service nova] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Refreshing network info cache for port f3d10573-dcc2-4f2b-b5bd-09bddb6d75d7 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 763.768167] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-51058cc8-1105-44e6-a426-742bbf5efe61 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.779755] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1818fd08-ebcf-434d-a0dc-c23e07d9393b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.811695] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 864ce643-8620-4405-b345-9164c81fec2a could not be found. [ 763.811695] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 763.812223] env[62965]: INFO nova.compute.manager [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Took 0.05 seconds to destroy the instance on the hypervisor. [ 763.812506] env[62965]: DEBUG oslo.service.loopingcall [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 763.815463] env[62965]: DEBUG nova.compute.manager [-] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 763.815584] env[62965]: DEBUG nova.network.neutron [-] [instance: 864ce643-8620-4405-b345-9164c81fec2a] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 763.842109] env[62965]: DEBUG nova.network.neutron [-] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 763.850424] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bb701a5-58f7-4656-84da-8802b8601d5a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.859542] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c052efc-e2a9-43e1-8ca3-70387b3c98c4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.890987] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-256a31b9-a60a-4662-8170-1700cc80832b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.897850] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59d48725-3823-4329-8b6c-fd5184647c96 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.911134] env[62965]: DEBUG nova.compute.provider_tree [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 764.132014] env[62965]: ERROR nova.compute.manager [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b60bf367-f412-4721-bb9f-2112fa9042de, please check neutron logs for more information. [ 764.132014] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 764.132014] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 764.132014] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 764.132014] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 764.132014] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 764.132014] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 764.132014] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 764.132014] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 764.132014] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 764.132014] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 764.132014] env[62965]: ERROR nova.compute.manager raise self.value [ 764.132014] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 764.132014] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 764.132014] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 764.132014] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 764.132660] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 764.132660] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 764.132660] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b60bf367-f412-4721-bb9f-2112fa9042de, please check neutron logs for more information. [ 764.132660] env[62965]: ERROR nova.compute.manager [ 764.132660] env[62965]: Traceback (most recent call last): [ 764.132660] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 764.132660] env[62965]: listener.cb(fileno) [ 764.132660] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 764.132660] env[62965]: result = function(*args, **kwargs) [ 764.132660] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 764.132660] env[62965]: return func(*args, **kwargs) [ 764.132660] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 764.132660] env[62965]: raise e [ 764.132660] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 764.132660] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 764.132660] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 764.132660] env[62965]: created_port_ids = self._update_ports_for_instance( [ 764.132660] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 764.132660] env[62965]: with excutils.save_and_reraise_exception(): [ 764.132660] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 764.132660] env[62965]: self.force_reraise() [ 764.132660] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 764.132660] env[62965]: raise self.value [ 764.132660] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 764.132660] env[62965]: updated_port = self._update_port( [ 764.132660] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 764.132660] env[62965]: _ensure_no_port_binding_failure(port) [ 764.132660] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 764.132660] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 764.133571] env[62965]: nova.exception.PortBindingFailed: Binding failed for port b60bf367-f412-4721-bb9f-2112fa9042de, please check neutron logs for more information. [ 764.133571] env[62965]: Removing descriptor: 19 [ 764.288321] env[62965]: DEBUG nova.network.neutron [req-83c152f4-1a70-4157-9215-d7889fa13b9a req-457535cd-999b-42e5-b589-669d9f1c0057 service nova] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 764.344831] env[62965]: DEBUG nova.network.neutron [-] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.346716] env[62965]: DEBUG nova.network.neutron [req-83c152f4-1a70-4157-9215-d7889fa13b9a req-457535cd-999b-42e5-b589-669d9f1c0057 service nova] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.413923] env[62965]: DEBUG nova.scheduler.client.report [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 764.545751] env[62965]: DEBUG nova.compute.manager [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 764.572182] env[62965]: DEBUG nova.virt.hardware [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 764.572432] env[62965]: DEBUG nova.virt.hardware [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 764.572586] env[62965]: DEBUG nova.virt.hardware [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 764.572764] env[62965]: DEBUG nova.virt.hardware [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 764.572906] env[62965]: DEBUG nova.virt.hardware [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 764.573063] env[62965]: DEBUG nova.virt.hardware [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 764.573268] env[62965]: DEBUG nova.virt.hardware [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 764.573423] env[62965]: DEBUG nova.virt.hardware [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 764.573585] env[62965]: DEBUG nova.virt.hardware [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 764.573745] env[62965]: DEBUG nova.virt.hardware [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 764.573916] env[62965]: DEBUG nova.virt.hardware [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 764.574781] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55efb051-51b5-47c3-9f50-61fdeaf6be56 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.582605] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3007a83-69db-4db9-93b3-91b9e9aa112a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.596144] env[62965]: ERROR nova.compute.manager [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b60bf367-f412-4721-bb9f-2112fa9042de, please check neutron logs for more information. [ 764.596144] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Traceback (most recent call last): [ 764.596144] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 764.596144] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] yield resources [ 764.596144] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 764.596144] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] self.driver.spawn(context, instance, image_meta, [ 764.596144] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 764.596144] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] self._vmops.spawn(context, instance, image_meta, injected_files, [ 764.596144] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 764.596144] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] vm_ref = self.build_virtual_machine(instance, [ 764.596144] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 764.596569] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] vif_infos = vmwarevif.get_vif_info(self._session, [ 764.596569] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 764.596569] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] for vif in network_info: [ 764.596569] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 764.596569] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] return self._sync_wrapper(fn, *args, **kwargs) [ 764.596569] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 764.596569] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] self.wait() [ 764.596569] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 764.596569] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] self[:] = self._gt.wait() [ 764.596569] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 764.596569] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] return self._exit_event.wait() [ 764.596569] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 764.596569] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] current.throw(*self._exc) [ 764.597333] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 764.597333] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] result = function(*args, **kwargs) [ 764.597333] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 764.597333] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] return func(*args, **kwargs) [ 764.597333] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 764.597333] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] raise e [ 764.597333] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 764.597333] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] nwinfo = self.network_api.allocate_for_instance( [ 764.597333] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 764.597333] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] created_port_ids = self._update_ports_for_instance( [ 764.597333] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 764.597333] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] with excutils.save_and_reraise_exception(): [ 764.597333] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 764.597914] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] self.force_reraise() [ 764.597914] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 764.597914] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] raise self.value [ 764.597914] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 764.597914] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] updated_port = self._update_port( [ 764.597914] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 764.597914] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] _ensure_no_port_binding_failure(port) [ 764.597914] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 764.597914] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] raise exception.PortBindingFailed(port_id=port['id']) [ 764.597914] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] nova.exception.PortBindingFailed: Binding failed for port b60bf367-f412-4721-bb9f-2112fa9042de, please check neutron logs for more information. [ 764.597914] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] [ 764.597914] env[62965]: INFO nova.compute.manager [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Terminating instance [ 764.850049] env[62965]: INFO nova.compute.manager [-] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Took 1.03 seconds to deallocate network for instance. [ 764.850049] env[62965]: DEBUG oslo_concurrency.lockutils [req-83c152f4-1a70-4157-9215-d7889fa13b9a req-457535cd-999b-42e5-b589-669d9f1c0057 service nova] Releasing lock "refresh_cache-864ce643-8620-4405-b345-9164c81fec2a" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.851531] env[62965]: DEBUG nova.compute.manager [req-83c152f4-1a70-4157-9215-d7889fa13b9a req-457535cd-999b-42e5-b589-669d9f1c0057 service nova] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Received event network-vif-deleted-f3d10573-dcc2-4f2b-b5bd-09bddb6d75d7 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 764.852655] env[62965]: DEBUG nova.compute.claims [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 764.852859] env[62965]: DEBUG oslo_concurrency.lockutils [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.919378] env[62965]: DEBUG oslo_concurrency.lockutils [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.395s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.919893] env[62965]: DEBUG nova.compute.manager [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: afaec369-6aef-4bdc-b886-eed795644867] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 764.922420] env[62965]: DEBUG oslo_concurrency.lockutils [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.620s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.923787] env[62965]: INFO nova.compute.claims [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 765.100469] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Acquiring lock "refresh_cache-d6e90043-ebd3-4daf-acfa-545ce8462942" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.100658] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Acquired lock "refresh_cache-d6e90043-ebd3-4daf-acfa-545ce8462942" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.100837] env[62965]: DEBUG nova.network.neutron [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 765.428367] env[62965]: DEBUG nova.compute.utils [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 765.431557] env[62965]: DEBUG nova.compute.manager [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: afaec369-6aef-4bdc-b886-eed795644867] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 765.431722] env[62965]: DEBUG nova.network.neutron [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: afaec369-6aef-4bdc-b886-eed795644867] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 765.469622] env[62965]: DEBUG nova.policy [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8cc49c20f2d549f587cca42df25962f8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd7298fbffd9d491e9bb59440f5889521', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 765.528340] env[62965]: DEBUG nova.compute.manager [req-60e2e4f8-cc6b-4798-a2f6-d78142a8b14c req-045dec42-cc59-47fc-8483-1fe3903a042f service nova] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Received event network-changed-b60bf367-f412-4721-bb9f-2112fa9042de {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 765.528524] env[62965]: DEBUG nova.compute.manager [req-60e2e4f8-cc6b-4798-a2f6-d78142a8b14c req-045dec42-cc59-47fc-8483-1fe3903a042f service nova] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Refreshing instance network info cache due to event network-changed-b60bf367-f412-4721-bb9f-2112fa9042de. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 765.528711] env[62965]: DEBUG oslo_concurrency.lockutils [req-60e2e4f8-cc6b-4798-a2f6-d78142a8b14c req-045dec42-cc59-47fc-8483-1fe3903a042f service nova] Acquiring lock "refresh_cache-d6e90043-ebd3-4daf-acfa-545ce8462942" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.617973] env[62965]: DEBUG nova.network.neutron [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 765.688015] env[62965]: DEBUG nova.network.neutron [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.729564] env[62965]: DEBUG nova.network.neutron [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: afaec369-6aef-4bdc-b886-eed795644867] Successfully created port: 316fb176-0010-4e0d-8df1-d84208f52dab {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 765.932665] env[62965]: DEBUG nova.compute.manager [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: afaec369-6aef-4bdc-b886-eed795644867] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 766.190967] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Releasing lock "refresh_cache-d6e90043-ebd3-4daf-acfa-545ce8462942" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.191708] env[62965]: DEBUG nova.compute.manager [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 766.191708] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 766.191959] env[62965]: DEBUG oslo_concurrency.lockutils [req-60e2e4f8-cc6b-4798-a2f6-d78142a8b14c req-045dec42-cc59-47fc-8483-1fe3903a042f service nova] Acquired lock "refresh_cache-d6e90043-ebd3-4daf-acfa-545ce8462942" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.192705] env[62965]: DEBUG nova.network.neutron [req-60e2e4f8-cc6b-4798-a2f6-d78142a8b14c req-045dec42-cc59-47fc-8483-1fe3903a042f service nova] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Refreshing network info cache for port b60bf367-f412-4721-bb9f-2112fa9042de {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 766.193172] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c54614a1-8830-402d-9dd8-0f74e2b18df0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.208020] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2d554b8-2f3f-46ed-a3ac-3082d0dd287f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.234379] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d6e90043-ebd3-4daf-acfa-545ce8462942 could not be found. [ 766.235499] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 766.235719] env[62965]: INFO nova.compute.manager [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Took 0.04 seconds to destroy the instance on the hypervisor. [ 766.235966] env[62965]: DEBUG oslo.service.loopingcall [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 766.238763] env[62965]: DEBUG nova.compute.manager [-] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 766.238947] env[62965]: DEBUG nova.network.neutron [-] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 766.264052] env[62965]: DEBUG nova.network.neutron [-] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 766.295181] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8984458c-48ac-4dd9-a738-b86889799c54 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.303137] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28fa2ecf-0762-4b8f-852e-40e166a63d60 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.335594] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0f6e55c-6a6d-4e3e-9a1e-c2f0ecde82a9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.346019] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10f7b161-aca7-479d-8343-6cbe5dd3c8cf {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.357310] env[62965]: DEBUG nova.compute.provider_tree [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 766.554349] env[62965]: ERROR nova.compute.manager [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 316fb176-0010-4e0d-8df1-d84208f52dab, please check neutron logs for more information. [ 766.554349] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 766.554349] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 766.554349] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 766.554349] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 766.554349] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 766.554349] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 766.554349] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 766.554349] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 766.554349] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 766.554349] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 766.554349] env[62965]: ERROR nova.compute.manager raise self.value [ 766.554349] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 766.554349] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 766.554349] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 766.554349] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 766.554916] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 766.554916] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 766.554916] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 316fb176-0010-4e0d-8df1-d84208f52dab, please check neutron logs for more information. [ 766.554916] env[62965]: ERROR nova.compute.manager [ 766.554916] env[62965]: Traceback (most recent call last): [ 766.554916] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 766.554916] env[62965]: listener.cb(fileno) [ 766.554916] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 766.554916] env[62965]: result = function(*args, **kwargs) [ 766.554916] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 766.554916] env[62965]: return func(*args, **kwargs) [ 766.554916] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 766.554916] env[62965]: raise e [ 766.554916] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 766.554916] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 766.554916] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 766.554916] env[62965]: created_port_ids = self._update_ports_for_instance( [ 766.554916] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 766.554916] env[62965]: with excutils.save_and_reraise_exception(): [ 766.554916] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 766.554916] env[62965]: self.force_reraise() [ 766.554916] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 766.554916] env[62965]: raise self.value [ 766.554916] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 766.554916] env[62965]: updated_port = self._update_port( [ 766.554916] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 766.554916] env[62965]: _ensure_no_port_binding_failure(port) [ 766.554916] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 766.554916] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 766.555918] env[62965]: nova.exception.PortBindingFailed: Binding failed for port 316fb176-0010-4e0d-8df1-d84208f52dab, please check neutron logs for more information. [ 766.555918] env[62965]: Removing descriptor: 19 [ 766.711786] env[62965]: DEBUG nova.network.neutron [req-60e2e4f8-cc6b-4798-a2f6-d78142a8b14c req-045dec42-cc59-47fc-8483-1fe3903a042f service nova] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 766.761714] env[62965]: DEBUG nova.network.neutron [req-60e2e4f8-cc6b-4798-a2f6-d78142a8b14c req-045dec42-cc59-47fc-8483-1fe3903a042f service nova] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.770382] env[62965]: DEBUG nova.network.neutron [-] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.861990] env[62965]: DEBUG nova.scheduler.client.report [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 766.943626] env[62965]: DEBUG nova.compute.manager [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: afaec369-6aef-4bdc-b886-eed795644867] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 766.970571] env[62965]: DEBUG nova.virt.hardware [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 766.970946] env[62965]: DEBUG nova.virt.hardware [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 766.971182] env[62965]: DEBUG nova.virt.hardware [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 766.971438] env[62965]: DEBUG nova.virt.hardware [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 766.971641] env[62965]: DEBUG nova.virt.hardware [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 766.971858] env[62965]: DEBUG nova.virt.hardware [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 766.972180] env[62965]: DEBUG nova.virt.hardware [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 766.972444] env[62965]: DEBUG nova.virt.hardware [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 766.972676] env[62965]: DEBUG nova.virt.hardware [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 766.972859] env[62965]: DEBUG nova.virt.hardware [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 766.973050] env[62965]: DEBUG nova.virt.hardware [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 766.973953] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad4a05a1-f03c-4e58-9ead-6166da47fb22 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.982060] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-131b5709-0dd2-4d70-bc8d-cea52b64250d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.995335] env[62965]: ERROR nova.compute.manager [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: afaec369-6aef-4bdc-b886-eed795644867] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 316fb176-0010-4e0d-8df1-d84208f52dab, please check neutron logs for more information. [ 766.995335] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] Traceback (most recent call last): [ 766.995335] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 766.995335] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] yield resources [ 766.995335] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 766.995335] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] self.driver.spawn(context, instance, image_meta, [ 766.995335] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 766.995335] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] self._vmops.spawn(context, instance, image_meta, injected_files, [ 766.995335] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 766.995335] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] vm_ref = self.build_virtual_machine(instance, [ 766.995335] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 766.995828] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] vif_infos = vmwarevif.get_vif_info(self._session, [ 766.995828] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 766.995828] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] for vif in network_info: [ 766.995828] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 766.995828] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] return self._sync_wrapper(fn, *args, **kwargs) [ 766.995828] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 766.995828] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] self.wait() [ 766.995828] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 766.995828] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] self[:] = self._gt.wait() [ 766.995828] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 766.995828] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] return self._exit_event.wait() [ 766.995828] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 766.995828] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] current.throw(*self._exc) [ 766.996332] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 766.996332] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] result = function(*args, **kwargs) [ 766.996332] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 766.996332] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] return func(*args, **kwargs) [ 766.996332] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 766.996332] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] raise e [ 766.996332] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 766.996332] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] nwinfo = self.network_api.allocate_for_instance( [ 766.996332] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 766.996332] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] created_port_ids = self._update_ports_for_instance( [ 766.996332] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 766.996332] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] with excutils.save_and_reraise_exception(): [ 766.996332] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 766.996780] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] self.force_reraise() [ 766.996780] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 766.996780] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] raise self.value [ 766.996780] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 766.996780] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] updated_port = self._update_port( [ 766.996780] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 766.996780] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] _ensure_no_port_binding_failure(port) [ 766.996780] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 766.996780] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] raise exception.PortBindingFailed(port_id=port['id']) [ 766.996780] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] nova.exception.PortBindingFailed: Binding failed for port 316fb176-0010-4e0d-8df1-d84208f52dab, please check neutron logs for more information. [ 766.996780] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] [ 766.996780] env[62965]: INFO nova.compute.manager [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: afaec369-6aef-4bdc-b886-eed795644867] Terminating instance [ 767.264471] env[62965]: DEBUG oslo_concurrency.lockutils [req-60e2e4f8-cc6b-4798-a2f6-d78142a8b14c req-045dec42-cc59-47fc-8483-1fe3903a042f service nova] Releasing lock "refresh_cache-d6e90043-ebd3-4daf-acfa-545ce8462942" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.264744] env[62965]: DEBUG nova.compute.manager [req-60e2e4f8-cc6b-4798-a2f6-d78142a8b14c req-045dec42-cc59-47fc-8483-1fe3903a042f service nova] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Received event network-vif-deleted-b60bf367-f412-4721-bb9f-2112fa9042de {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 767.274497] env[62965]: INFO nova.compute.manager [-] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Took 1.04 seconds to deallocate network for instance. [ 767.279101] env[62965]: DEBUG nova.compute.claims [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 767.279299] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.367718] env[62965]: DEBUG oslo_concurrency.lockutils [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.445s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.368644] env[62965]: DEBUG nova.compute.manager [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 767.371022] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 16.268s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.371022] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.371154] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62965) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 767.372412] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.100s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.372796] env[62965]: INFO nova.compute.claims [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 767.376339] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-746306da-76dc-4ddc-8aa6-d09705a06443 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.384661] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb4c06eb-a4d2-4526-8985-e0166248981f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.399899] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-391ac987-060e-4535-823f-084cf53e2492 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.407282] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7a8986e-95b1-4c36-bc3f-c99542693c09 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.437600] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181383MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62965) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 767.437760] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.499662] env[62965]: DEBUG oslo_concurrency.lockutils [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Acquiring lock "refresh_cache-afaec369-6aef-4bdc-b886-eed795644867" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.499885] env[62965]: DEBUG oslo_concurrency.lockutils [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Acquired lock "refresh_cache-afaec369-6aef-4bdc-b886-eed795644867" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.500098] env[62965]: DEBUG nova.network.neutron [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: afaec369-6aef-4bdc-b886-eed795644867] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 767.552105] env[62965]: DEBUG nova.compute.manager [req-4888232f-dd87-4db6-adc9-f5382794f454 req-00a1de05-0f63-4975-8cb5-742dfc36a20e service nova] [instance: afaec369-6aef-4bdc-b886-eed795644867] Received event network-changed-316fb176-0010-4e0d-8df1-d84208f52dab {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 767.552304] env[62965]: DEBUG nova.compute.manager [req-4888232f-dd87-4db6-adc9-f5382794f454 req-00a1de05-0f63-4975-8cb5-742dfc36a20e service nova] [instance: afaec369-6aef-4bdc-b886-eed795644867] Refreshing instance network info cache due to event network-changed-316fb176-0010-4e0d-8df1-d84208f52dab. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 767.552498] env[62965]: DEBUG oslo_concurrency.lockutils [req-4888232f-dd87-4db6-adc9-f5382794f454 req-00a1de05-0f63-4975-8cb5-742dfc36a20e service nova] Acquiring lock "refresh_cache-afaec369-6aef-4bdc-b886-eed795644867" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.877879] env[62965]: DEBUG nova.compute.utils [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 767.879376] env[62965]: DEBUG nova.compute.manager [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Not allocating networking since 'none' was specified. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 768.022224] env[62965]: DEBUG nova.network.neutron [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: afaec369-6aef-4bdc-b886-eed795644867] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 768.101995] env[62965]: DEBUG nova.network.neutron [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: afaec369-6aef-4bdc-b886-eed795644867] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.383900] env[62965]: DEBUG nova.compute.manager [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 768.605297] env[62965]: DEBUG oslo_concurrency.lockutils [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Releasing lock "refresh_cache-afaec369-6aef-4bdc-b886-eed795644867" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.605819] env[62965]: DEBUG nova.compute.manager [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: afaec369-6aef-4bdc-b886-eed795644867] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 768.605989] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: afaec369-6aef-4bdc-b886-eed795644867] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 768.606493] env[62965]: DEBUG oslo_concurrency.lockutils [req-4888232f-dd87-4db6-adc9-f5382794f454 req-00a1de05-0f63-4975-8cb5-742dfc36a20e service nova] Acquired lock "refresh_cache-afaec369-6aef-4bdc-b886-eed795644867" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.606674] env[62965]: DEBUG nova.network.neutron [req-4888232f-dd87-4db6-adc9-f5382794f454 req-00a1de05-0f63-4975-8cb5-742dfc36a20e service nova] [instance: afaec369-6aef-4bdc-b886-eed795644867] Refreshing network info cache for port 316fb176-0010-4e0d-8df1-d84208f52dab {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 768.607771] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b2cee0eb-98ec-4dc2-bd84-43336a67d6d7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.616535] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86ca4a8a-1d2c-4493-bf9b-e9117786d9f4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.641879] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: afaec369-6aef-4bdc-b886-eed795644867] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance afaec369-6aef-4bdc-b886-eed795644867 could not be found. [ 768.642452] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: afaec369-6aef-4bdc-b886-eed795644867] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 768.642651] env[62965]: INFO nova.compute.manager [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: afaec369-6aef-4bdc-b886-eed795644867] Took 0.04 seconds to destroy the instance on the hypervisor. [ 768.642898] env[62965]: DEBUG oslo.service.loopingcall [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 768.646281] env[62965]: DEBUG nova.compute.manager [-] [instance: afaec369-6aef-4bdc-b886-eed795644867] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 768.646281] env[62965]: DEBUG nova.network.neutron [-] [instance: afaec369-6aef-4bdc-b886-eed795644867] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 768.661311] env[62965]: DEBUG nova.network.neutron [-] [instance: afaec369-6aef-4bdc-b886-eed795644867] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 768.670797] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-533fac2f-cbcb-4331-9522-e4f96982111a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.678614] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d418174-195e-4588-a961-bd7c2247384e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.709815] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e11b326-a5bf-472e-bbcc-51e641da8f06 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.716939] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6228a46-e1c4-4bcb-b124-5f4afbed09a7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.730458] env[62965]: DEBUG nova.compute.provider_tree [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 769.126733] env[62965]: DEBUG nova.network.neutron [req-4888232f-dd87-4db6-adc9-f5382794f454 req-00a1de05-0f63-4975-8cb5-742dfc36a20e service nova] [instance: afaec369-6aef-4bdc-b886-eed795644867] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 769.163886] env[62965]: DEBUG nova.network.neutron [-] [instance: afaec369-6aef-4bdc-b886-eed795644867] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.192308] env[62965]: DEBUG nova.network.neutron [req-4888232f-dd87-4db6-adc9-f5382794f454 req-00a1de05-0f63-4975-8cb5-742dfc36a20e service nova] [instance: afaec369-6aef-4bdc-b886-eed795644867] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.234470] env[62965]: DEBUG nova.scheduler.client.report [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 769.395840] env[62965]: DEBUG nova.compute.manager [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 769.417136] env[62965]: DEBUG nova.virt.hardware [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 769.417395] env[62965]: DEBUG nova.virt.hardware [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 769.417549] env[62965]: DEBUG nova.virt.hardware [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 769.417726] env[62965]: DEBUG nova.virt.hardware [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 769.417871] env[62965]: DEBUG nova.virt.hardware [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 769.418019] env[62965]: DEBUG nova.virt.hardware [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 769.418291] env[62965]: DEBUG nova.virt.hardware [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 769.418460] env[62965]: DEBUG nova.virt.hardware [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 769.418625] env[62965]: DEBUG nova.virt.hardware [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 769.418785] env[62965]: DEBUG nova.virt.hardware [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 769.419047] env[62965]: DEBUG nova.virt.hardware [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 769.419976] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08c7638d-ad3c-4a31-8479-f638dd6af0c8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.428230] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cffa2449-c83d-48db-b320-99ab1118a30f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.442198] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Instance VIF info [] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 769.448185] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Creating folder: Project (eccd39df00d2403f929981d7f0613edd). Parent ref: group-v294931. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 769.448457] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c9af8691-5238-48ce-bf24-68adeec6c17d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.459203] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Created folder: Project (eccd39df00d2403f929981d7f0613edd) in parent group-v294931. [ 769.459422] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Creating folder: Instances. Parent ref: group-v294948. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 769.459567] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-85818e7a-ea18-462a-b124-90266e9814ce {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.468309] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Created folder: Instances in parent group-v294948. [ 769.469022] env[62965]: DEBUG oslo.service.loopingcall [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 769.469022] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 769.469267] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1eca17c0-7a0d-4d2b-8348-3197363fb96a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.484867] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 769.484867] env[62965]: value = "task-1389787" [ 769.484867] env[62965]: _type = "Task" [ 769.484867] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.492460] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389787, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.666361] env[62965]: INFO nova.compute.manager [-] [instance: afaec369-6aef-4bdc-b886-eed795644867] Took 1.02 seconds to deallocate network for instance. [ 769.668708] env[62965]: DEBUG nova.compute.claims [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: afaec369-6aef-4bdc-b886-eed795644867] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 769.668916] env[62965]: DEBUG oslo_concurrency.lockutils [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.694687] env[62965]: DEBUG oslo_concurrency.lockutils [req-4888232f-dd87-4db6-adc9-f5382794f454 req-00a1de05-0f63-4975-8cb5-742dfc36a20e service nova] Releasing lock "refresh_cache-afaec369-6aef-4bdc-b886-eed795644867" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.694930] env[62965]: DEBUG nova.compute.manager [req-4888232f-dd87-4db6-adc9-f5382794f454 req-00a1de05-0f63-4975-8cb5-742dfc36a20e service nova] [instance: afaec369-6aef-4bdc-b886-eed795644867] Received event network-vif-deleted-316fb176-0010-4e0d-8df1-d84208f52dab {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 769.739174] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.367s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.739687] env[62965]: DEBUG nova.compute.manager [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 769.742328] env[62965]: DEBUG oslo_concurrency.lockutils [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.524s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.743687] env[62965]: INFO nova.compute.claims [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 769.995996] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389787, 'name': CreateVM_Task, 'duration_secs': 0.24011} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.996209] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 769.996636] env[62965]: DEBUG oslo_concurrency.lockutils [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.996799] env[62965]: DEBUG oslo_concurrency.lockutils [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.997163] env[62965]: DEBUG oslo_concurrency.lockutils [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 769.997427] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38eea4fc-f065-462f-a03a-be84aac3e6f4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.002287] env[62965]: DEBUG oslo_vmware.api [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Waiting for the task: (returnval){ [ 770.002287] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]523f9b37-d44b-7748-de0c-70ae356959aa" [ 770.002287] env[62965]: _type = "Task" [ 770.002287] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.015752] env[62965]: DEBUG oslo_vmware.api [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]523f9b37-d44b-7748-de0c-70ae356959aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.247733] env[62965]: DEBUG nova.compute.utils [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 770.251065] env[62965]: DEBUG nova.compute.manager [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Not allocating networking since 'none' was specified. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 770.513518] env[62965]: DEBUG oslo_vmware.api [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]523f9b37-d44b-7748-de0c-70ae356959aa, 'name': SearchDatastore_Task, 'duration_secs': 0.011496} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.514077] env[62965]: DEBUG oslo_concurrency.lockutils [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.514487] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 770.515752] env[62965]: DEBUG oslo_concurrency.lockutils [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.515752] env[62965]: DEBUG oslo_concurrency.lockutils [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.515752] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 770.515752] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-12df63eb-1dfd-4138-ad7d-e713990c7d46 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.523874] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 770.524099] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 770.524838] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d77b177b-13c9-4a01-be97-6f25c46459de {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.530272] env[62965]: DEBUG oslo_vmware.api [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Waiting for the task: (returnval){ [ 770.530272] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]520ecad8-e216-8837-ece3-1560734c2a34" [ 770.530272] env[62965]: _type = "Task" [ 770.530272] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.538576] env[62965]: DEBUG oslo_vmware.api [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]520ecad8-e216-8837-ece3-1560734c2a34, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.751717] env[62965]: DEBUG nova.compute.manager [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 770.985672] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dce4ad0-001b-407d-b166-3579bd96d4dd {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.993021] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2ed4909-5705-4646-b605-db205043ba2f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.021933] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f1682a1-4954-4aaa-8366-dfe64fd89b59 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.028600] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08d25e38-dd56-42bb-865c-9d16f121ac2b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.045075] env[62965]: DEBUG nova.compute.provider_tree [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 771.049617] env[62965]: DEBUG oslo_vmware.api [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]520ecad8-e216-8837-ece3-1560734c2a34, 'name': SearchDatastore_Task, 'duration_secs': 0.008937} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.050092] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-532a9872-6ea8-4132-b6ba-82c3d1c38aa5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.054716] env[62965]: DEBUG oslo_vmware.api [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Waiting for the task: (returnval){ [ 771.054716] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]529f44e6-6f5c-608a-7a5f-f4b5f7d98e1c" [ 771.054716] env[62965]: _type = "Task" [ 771.054716] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.061945] env[62965]: DEBUG oslo_vmware.api [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]529f44e6-6f5c-608a-7a5f-f4b5f7d98e1c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.551361] env[62965]: DEBUG nova.scheduler.client.report [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 771.564822] env[62965]: DEBUG oslo_vmware.api [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]529f44e6-6f5c-608a-7a5f-f4b5f7d98e1c, 'name': SearchDatastore_Task, 'duration_secs': 0.008061} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.565096] env[62965]: DEBUG oslo_concurrency.lockutils [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.565355] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 6e0a4e4b-21a2-4185-999e-69b4c74c37d7/6e0a4e4b-21a2-4185-999e-69b4c74c37d7.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 771.566251] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6ee17552-e828-4dd8-a6b0-a3eded2b0c32 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.573681] env[62965]: DEBUG oslo_vmware.api [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Waiting for the task: (returnval){ [ 771.573681] env[62965]: value = "task-1389788" [ 771.573681] env[62965]: _type = "Task" [ 771.573681] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.582129] env[62965]: DEBUG oslo_vmware.api [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389788, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.764209] env[62965]: DEBUG nova.compute.manager [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 771.788761] env[62965]: DEBUG nova.virt.hardware [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 771.789124] env[62965]: DEBUG nova.virt.hardware [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 771.789846] env[62965]: DEBUG nova.virt.hardware [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 771.789846] env[62965]: DEBUG nova.virt.hardware [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 771.789846] env[62965]: DEBUG nova.virt.hardware [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 771.790188] env[62965]: DEBUG nova.virt.hardware [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 771.790279] env[62965]: DEBUG nova.virt.hardware [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 771.790499] env[62965]: DEBUG nova.virt.hardware [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 771.790742] env[62965]: DEBUG nova.virt.hardware [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 771.791287] env[62965]: DEBUG nova.virt.hardware [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 771.791287] env[62965]: DEBUG nova.virt.hardware [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 771.792600] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c921c298-ce8e-4d14-bca7-6ecb895b3e08 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.802269] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8d118c2-bbbc-4821-b246-0203c82642e3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.817783] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Instance VIF info [] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 771.823790] env[62965]: DEBUG oslo.service.loopingcall [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 771.824224] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 771.824367] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-27bbfd21-438e-4a32-9856-2a4b81925da5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.844488] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 771.844488] env[62965]: value = "task-1389789" [ 771.844488] env[62965]: _type = "Task" [ 771.844488] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.854578] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389789, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.057251] env[62965]: DEBUG oslo_concurrency.lockutils [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.315s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.057939] env[62965]: DEBUG nova.compute.manager [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 772.061833] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.299s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.084280] env[62965]: DEBUG oslo_vmware.api [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389788, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.506165} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.085173] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 6e0a4e4b-21a2-4185-999e-69b4c74c37d7/6e0a4e4b-21a2-4185-999e-69b4c74c37d7.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 772.085173] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 772.085173] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fc9ae200-6748-440c-94be-bd8e14463ca9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.091157] env[62965]: DEBUG oslo_vmware.api [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Waiting for the task: (returnval){ [ 772.091157] env[62965]: value = "task-1389790" [ 772.091157] env[62965]: _type = "Task" [ 772.091157] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.100180] env[62965]: DEBUG oslo_vmware.api [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389790, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.357092] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389789, 'name': CreateVM_Task, 'duration_secs': 0.426816} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.357092] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 772.357365] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.357557] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.357872] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 772.358348] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a0f01c8-8007-43bf-ac3a-638847afecd5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.363759] env[62965]: DEBUG oslo_vmware.api [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Waiting for the task: (returnval){ [ 772.363759] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]520206f6-0695-1777-c258-a721860448ea" [ 772.363759] env[62965]: _type = "Task" [ 772.363759] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.372076] env[62965]: DEBUG oslo_vmware.api [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]520206f6-0695-1777-c258-a721860448ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.566823] env[62965]: DEBUG nova.compute.utils [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 772.571756] env[62965]: DEBUG nova.compute.manager [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 772.572143] env[62965]: DEBUG nova.network.neutron [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 772.601529] env[62965]: DEBUG oslo_vmware.api [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389790, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.103694} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.604099] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 772.605152] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9018b2a8-083b-4a2a-b6c8-2c104b2d926a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.625728] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Reconfiguring VM instance instance-00000031 to attach disk [datastore2] 6e0a4e4b-21a2-4185-999e-69b4c74c37d7/6e0a4e4b-21a2-4185-999e-69b4c74c37d7.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 772.628542] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-af6f65ff-d292-4bb9-b289-35f171de8ace {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.643917] env[62965]: DEBUG nova.policy [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '44a4caf0121444529af77b85f01de8c0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5066ce28a2c24bfaa311966545306b8f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 772.651706] env[62965]: DEBUG oslo_vmware.api [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Waiting for the task: (returnval){ [ 772.651706] env[62965]: value = "task-1389791" [ 772.651706] env[62965]: _type = "Task" [ 772.651706] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.662684] env[62965]: DEBUG oslo_vmware.api [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389791, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.875620] env[62965]: DEBUG oslo_vmware.api [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]520206f6-0695-1777-c258-a721860448ea, 'name': SearchDatastore_Task, 'duration_secs': 0.009354} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.875928] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.876179] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 772.876415] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.876559] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.876733] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 772.877084] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e084e9bc-fcec-4cd1-acc5-6bd191a05ac6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.891838] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 772.892088] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 772.893275] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b142e338-bba4-4d81-871e-6f8b20985de8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.900328] env[62965]: DEBUG oslo_vmware.api [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Waiting for the task: (returnval){ [ 772.900328] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]520df13d-019f-65b2-1a92-533f0b2af212" [ 772.900328] env[62965]: _type = "Task" [ 772.900328] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.908419] env[62965]: DEBUG oslo_vmware.api [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]520df13d-019f-65b2-1a92-533f0b2af212, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.922436] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6bb927c-bbd7-4857-abb2-cd2c5a530638 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.929317] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-404041cf-f0d6-49a0-824e-07c58149a3b8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.961785] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-939071f2-ac92-44bd-89f9-c9c93b8111a9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.969235] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33e75308-78ea-4c1b-85cb-06b53b081a15 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.982841] env[62965]: DEBUG nova.compute.provider_tree [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 772.985115] env[62965]: DEBUG nova.network.neutron [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Successfully created port: 6f417bf0-8e7f-4fa8-96c5-e21e200204c3 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 773.073189] env[62965]: DEBUG nova.compute.manager [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 773.162772] env[62965]: DEBUG oslo_vmware.api [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389791, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.410941] env[62965]: DEBUG oslo_vmware.api [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]520df13d-019f-65b2-1a92-533f0b2af212, 'name': SearchDatastore_Task, 'duration_secs': 0.01135} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.412261] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-330ef6e7-851e-40ff-8ee5-1f1bdf25856d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.417861] env[62965]: DEBUG oslo_vmware.api [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Waiting for the task: (returnval){ [ 773.417861] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52b54481-3934-82fc-0112-dfa8a7f3b1c2" [ 773.417861] env[62965]: _type = "Task" [ 773.417861] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.426199] env[62965]: DEBUG oslo_vmware.api [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52b54481-3934-82fc-0112-dfa8a7f3b1c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.488535] env[62965]: DEBUG nova.scheduler.client.report [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 773.646240] env[62965]: DEBUG nova.compute.manager [req-cb887ef0-94fc-49b3-8185-85fc4551e4ac req-6b862be3-18aa-41c6-b944-c9ebc4e7838f service nova] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Received event network-changed-6f417bf0-8e7f-4fa8-96c5-e21e200204c3 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 773.646240] env[62965]: DEBUG nova.compute.manager [req-cb887ef0-94fc-49b3-8185-85fc4551e4ac req-6b862be3-18aa-41c6-b944-c9ebc4e7838f service nova] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Refreshing instance network info cache due to event network-changed-6f417bf0-8e7f-4fa8-96c5-e21e200204c3. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 773.646487] env[62965]: DEBUG oslo_concurrency.lockutils [req-cb887ef0-94fc-49b3-8185-85fc4551e4ac req-6b862be3-18aa-41c6-b944-c9ebc4e7838f service nova] Acquiring lock "refresh_cache-9a2a324c-4eb1-45b6-8414-42671b412429" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.646581] env[62965]: DEBUG oslo_concurrency.lockutils [req-cb887ef0-94fc-49b3-8185-85fc4551e4ac req-6b862be3-18aa-41c6-b944-c9ebc4e7838f service nova] Acquired lock "refresh_cache-9a2a324c-4eb1-45b6-8414-42671b412429" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.646674] env[62965]: DEBUG nova.network.neutron [req-cb887ef0-94fc-49b3-8185-85fc4551e4ac req-6b862be3-18aa-41c6-b944-c9ebc4e7838f service nova] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Refreshing network info cache for port 6f417bf0-8e7f-4fa8-96c5-e21e200204c3 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 773.664495] env[62965]: DEBUG oslo_vmware.api [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389791, 'name': ReconfigVM_Task, 'duration_secs': 0.6237} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.664726] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Reconfigured VM instance instance-00000031 to attach disk [datastore2] 6e0a4e4b-21a2-4185-999e-69b4c74c37d7/6e0a4e4b-21a2-4185-999e-69b4c74c37d7.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 773.665359] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b1cab66f-57ef-47ea-9754-55a96559ec5c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.673336] env[62965]: DEBUG oslo_vmware.api [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Waiting for the task: (returnval){ [ 773.673336] env[62965]: value = "task-1389792" [ 773.673336] env[62965]: _type = "Task" [ 773.673336] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.682328] env[62965]: DEBUG oslo_vmware.api [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389792, 'name': Rename_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.838672] env[62965]: ERROR nova.compute.manager [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6f417bf0-8e7f-4fa8-96c5-e21e200204c3, please check neutron logs for more information. [ 773.838672] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 773.838672] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 773.838672] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 773.838672] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 773.838672] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 773.838672] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 773.838672] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 773.838672] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 773.838672] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 773.838672] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 773.838672] env[62965]: ERROR nova.compute.manager raise self.value [ 773.838672] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 773.838672] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 773.838672] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 773.838672] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 773.839388] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 773.839388] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 773.839388] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6f417bf0-8e7f-4fa8-96c5-e21e200204c3, please check neutron logs for more information. [ 773.839388] env[62965]: ERROR nova.compute.manager [ 773.839388] env[62965]: Traceback (most recent call last): [ 773.839388] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 773.839388] env[62965]: listener.cb(fileno) [ 773.839388] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 773.839388] env[62965]: result = function(*args, **kwargs) [ 773.839388] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 773.839388] env[62965]: return func(*args, **kwargs) [ 773.839388] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 773.839388] env[62965]: raise e [ 773.839388] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 773.839388] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 773.839388] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 773.839388] env[62965]: created_port_ids = self._update_ports_for_instance( [ 773.839388] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 773.839388] env[62965]: with excutils.save_and_reraise_exception(): [ 773.839388] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 773.839388] env[62965]: self.force_reraise() [ 773.839388] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 773.839388] env[62965]: raise self.value [ 773.839388] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 773.839388] env[62965]: updated_port = self._update_port( [ 773.839388] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 773.839388] env[62965]: _ensure_no_port_binding_failure(port) [ 773.839388] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 773.839388] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 773.840453] env[62965]: nova.exception.PortBindingFailed: Binding failed for port 6f417bf0-8e7f-4fa8-96c5-e21e200204c3, please check neutron logs for more information. [ 773.840453] env[62965]: Removing descriptor: 19 [ 773.928502] env[62965]: DEBUG oslo_vmware.api [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52b54481-3934-82fc-0112-dfa8a7f3b1c2, 'name': SearchDatastore_Task, 'duration_secs': 0.009616} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.928758] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.929048] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] e87adfae-3ea4-4cc4-9978-d8c811c80941/e87adfae-3ea4-4cc4-9978-d8c811c80941.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 773.929305] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a7cf2aa4-986b-4e43-bf7a-e07af54834eb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.936027] env[62965]: DEBUG oslo_vmware.api [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Waiting for the task: (returnval){ [ 773.936027] env[62965]: value = "task-1389793" [ 773.936027] env[62965]: _type = "Task" [ 773.936027] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.942884] env[62965]: DEBUG oslo_vmware.api [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389793, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.994117] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.932s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.994797] env[62965]: ERROR nova.compute.manager [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 171591e5-b76f-4852-b993-0c69d8392a0c, please check neutron logs for more information. [ 773.994797] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Traceback (most recent call last): [ 773.994797] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 773.994797] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] self.driver.spawn(context, instance, image_meta, [ 773.994797] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 773.994797] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] self._vmops.spawn(context, instance, image_meta, injected_files, [ 773.994797] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 773.994797] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] vm_ref = self.build_virtual_machine(instance, [ 773.994797] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 773.994797] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] vif_infos = vmwarevif.get_vif_info(self._session, [ 773.994797] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 773.995249] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] for vif in network_info: [ 773.995249] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 773.995249] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] return self._sync_wrapper(fn, *args, **kwargs) [ 773.995249] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 773.995249] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] self.wait() [ 773.995249] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 773.995249] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] self[:] = self._gt.wait() [ 773.995249] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 773.995249] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] return self._exit_event.wait() [ 773.995249] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 773.995249] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] result = hub.switch() [ 773.995249] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 773.995249] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] return self.greenlet.switch() [ 773.995751] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 773.995751] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] result = function(*args, **kwargs) [ 773.995751] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 773.995751] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] return func(*args, **kwargs) [ 773.995751] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 773.995751] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] raise e [ 773.995751] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 773.995751] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] nwinfo = self.network_api.allocate_for_instance( [ 773.995751] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 773.995751] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] created_port_ids = self._update_ports_for_instance( [ 773.995751] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 773.995751] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] with excutils.save_and_reraise_exception(): [ 773.995751] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 773.996195] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] self.force_reraise() [ 773.996195] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 773.996195] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] raise self.value [ 773.996195] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 773.996195] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] updated_port = self._update_port( [ 773.996195] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 773.996195] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] _ensure_no_port_binding_failure(port) [ 773.996195] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 773.996195] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] raise exception.PortBindingFailed(port_id=port['id']) [ 773.996195] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] nova.exception.PortBindingFailed: Binding failed for port 171591e5-b76f-4852-b993-0c69d8392a0c, please check neutron logs for more information. [ 773.996195] env[62965]: ERROR nova.compute.manager [instance: cd978d56-ba80-44a9-9a02-3a1254066571] [ 773.996596] env[62965]: DEBUG nova.compute.utils [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Binding failed for port 171591e5-b76f-4852-b993-0c69d8392a0c, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 773.996958] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.348s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.998522] env[62965]: INFO nova.compute.claims [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 774.001398] env[62965]: DEBUG nova.compute.manager [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Build of instance cd978d56-ba80-44a9-9a02-3a1254066571 was re-scheduled: Binding failed for port 171591e5-b76f-4852-b993-0c69d8392a0c, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 774.001824] env[62965]: DEBUG nova.compute.manager [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 774.002058] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] Acquiring lock "refresh_cache-cd978d56-ba80-44a9-9a02-3a1254066571" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.002212] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] Acquired lock "refresh_cache-cd978d56-ba80-44a9-9a02-3a1254066571" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.002369] env[62965]: DEBUG nova.network.neutron [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 774.082304] env[62965]: DEBUG nova.compute.manager [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 774.109281] env[62965]: DEBUG nova.virt.hardware [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 774.109522] env[62965]: DEBUG nova.virt.hardware [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 774.109675] env[62965]: DEBUG nova.virt.hardware [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 774.109848] env[62965]: DEBUG nova.virt.hardware [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 774.110179] env[62965]: DEBUG nova.virt.hardware [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 774.110381] env[62965]: DEBUG nova.virt.hardware [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 774.110588] env[62965]: DEBUG nova.virt.hardware [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 774.110744] env[62965]: DEBUG nova.virt.hardware [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 774.110908] env[62965]: DEBUG nova.virt.hardware [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 774.111079] env[62965]: DEBUG nova.virt.hardware [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 774.111254] env[62965]: DEBUG nova.virt.hardware [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 774.112124] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef491119-314b-4831-b887-52ada8c42da3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.120618] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63bbb1b3-63d9-4211-8680-e6560ab50e07 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.136721] env[62965]: ERROR nova.compute.manager [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6f417bf0-8e7f-4fa8-96c5-e21e200204c3, please check neutron logs for more information. [ 774.136721] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Traceback (most recent call last): [ 774.136721] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 774.136721] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] yield resources [ 774.136721] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 774.136721] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] self.driver.spawn(context, instance, image_meta, [ 774.136721] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 774.136721] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] self._vmops.spawn(context, instance, image_meta, injected_files, [ 774.136721] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 774.136721] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] vm_ref = self.build_virtual_machine(instance, [ 774.136721] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 774.137329] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] vif_infos = vmwarevif.get_vif_info(self._session, [ 774.137329] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 774.137329] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] for vif in network_info: [ 774.137329] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 774.137329] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] return self._sync_wrapper(fn, *args, **kwargs) [ 774.137329] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 774.137329] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] self.wait() [ 774.137329] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 774.137329] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] self[:] = self._gt.wait() [ 774.137329] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 774.137329] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] return self._exit_event.wait() [ 774.137329] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 774.137329] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] current.throw(*self._exc) [ 774.137823] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 774.137823] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] result = function(*args, **kwargs) [ 774.137823] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 774.137823] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] return func(*args, **kwargs) [ 774.137823] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 774.137823] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] raise e [ 774.137823] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 774.137823] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] nwinfo = self.network_api.allocate_for_instance( [ 774.137823] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 774.137823] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] created_port_ids = self._update_ports_for_instance( [ 774.137823] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 774.137823] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] with excutils.save_and_reraise_exception(): [ 774.137823] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 774.138747] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] self.force_reraise() [ 774.138747] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 774.138747] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] raise self.value [ 774.138747] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 774.138747] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] updated_port = self._update_port( [ 774.138747] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 774.138747] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] _ensure_no_port_binding_failure(port) [ 774.138747] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 774.138747] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] raise exception.PortBindingFailed(port_id=port['id']) [ 774.138747] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] nova.exception.PortBindingFailed: Binding failed for port 6f417bf0-8e7f-4fa8-96c5-e21e200204c3, please check neutron logs for more information. [ 774.138747] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] [ 774.138747] env[62965]: INFO nova.compute.manager [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Terminating instance [ 774.164012] env[62965]: DEBUG nova.network.neutron [req-cb887ef0-94fc-49b3-8185-85fc4551e4ac req-6b862be3-18aa-41c6-b944-c9ebc4e7838f service nova] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 774.185814] env[62965]: DEBUG oslo_vmware.api [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389792, 'name': Rename_Task, 'duration_secs': 0.149669} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.186100] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 774.186352] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-76c65b85-2dc9-4d2a-bf85-2f58c7bc7421 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.193848] env[62965]: DEBUG oslo_vmware.api [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Waiting for the task: (returnval){ [ 774.193848] env[62965]: value = "task-1389794" [ 774.193848] env[62965]: _type = "Task" [ 774.193848] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.210524] env[62965]: DEBUG oslo_vmware.api [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389794, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.241825] env[62965]: DEBUG nova.network.neutron [req-cb887ef0-94fc-49b3-8185-85fc4551e4ac req-6b862be3-18aa-41c6-b944-c9ebc4e7838f service nova] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.445421] env[62965]: DEBUG oslo_vmware.api [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389793, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.521109] env[62965]: DEBUG nova.network.neutron [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 774.573602] env[62965]: DEBUG nova.network.neutron [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.641268] env[62965]: DEBUG oslo_concurrency.lockutils [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Acquiring lock "refresh_cache-9a2a324c-4eb1-45b6-8414-42671b412429" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.704936] env[62965]: DEBUG oslo_vmware.api [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389794, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.745142] env[62965]: DEBUG oslo_concurrency.lockutils [req-cb887ef0-94fc-49b3-8185-85fc4551e4ac req-6b862be3-18aa-41c6-b944-c9ebc4e7838f service nova] Releasing lock "refresh_cache-9a2a324c-4eb1-45b6-8414-42671b412429" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.745578] env[62965]: DEBUG oslo_concurrency.lockutils [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Acquired lock "refresh_cache-9a2a324c-4eb1-45b6-8414-42671b412429" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.745767] env[62965]: DEBUG nova.network.neutron [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 774.946950] env[62965]: DEBUG oslo_vmware.api [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389793, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.51361} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.947586] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] e87adfae-3ea4-4cc4-9978-d8c811c80941/e87adfae-3ea4-4cc4-9978-d8c811c80941.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 774.950019] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 774.950019] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f47d2244-d422-4cf4-a86f-a3c35556a504 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.956036] env[62965]: DEBUG oslo_vmware.api [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Waiting for the task: (returnval){ [ 774.956036] env[62965]: value = "task-1389795" [ 774.956036] env[62965]: _type = "Task" [ 774.956036] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.964467] env[62965]: DEBUG oslo_vmware.api [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389795, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.076231] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] Releasing lock "refresh_cache-cd978d56-ba80-44a9-9a02-3a1254066571" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.076463] env[62965]: DEBUG nova.compute.manager [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 775.076622] env[62965]: DEBUG nova.compute.manager [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 775.076803] env[62965]: DEBUG nova.network.neutron [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 775.091360] env[62965]: DEBUG nova.network.neutron [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 775.206700] env[62965]: DEBUG oslo_vmware.api [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389794, 'name': PowerOnVM_Task, 'duration_secs': 0.665124} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.206952] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 775.207164] env[62965]: INFO nova.compute.manager [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Took 5.81 seconds to spawn the instance on the hypervisor. [ 775.207334] env[62965]: DEBUG nova.compute.manager [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 775.208070] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7caa6c99-2118-44e2-a43e-85397e0e6299 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.262616] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23efe2fd-f795-46a7-9328-3247087a9c03 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.265587] env[62965]: DEBUG nova.network.neutron [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 775.271766] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3fdf889-352a-468c-abdd-6444e1657b8e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.301365] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4f57c5d-d7fd-4229-b330-a79cffde2827 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.311154] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-048d53dd-9aae-4f76-8f81-d2aa0beae944 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.323960] env[62965]: DEBUG nova.compute.provider_tree [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 775.339340] env[62965]: DEBUG nova.network.neutron [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.466248] env[62965]: DEBUG oslo_vmware.api [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389795, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.121945} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.466572] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 775.467427] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feba85a1-addf-4558-bc76-28baab7ed518 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.487592] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Reconfiguring VM instance instance-00000032 to attach disk [datastore2] e87adfae-3ea4-4cc4-9978-d8c811c80941/e87adfae-3ea4-4cc4-9978-d8c811c80941.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 775.487864] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2a1d7b86-020f-4fd0-a697-8e000fb01b4d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.509031] env[62965]: DEBUG oslo_vmware.api [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Waiting for the task: (returnval){ [ 775.509031] env[62965]: value = "task-1389796" [ 775.509031] env[62965]: _type = "Task" [ 775.509031] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.515818] env[62965]: DEBUG oslo_vmware.api [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389796, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.593760] env[62965]: DEBUG nova.network.neutron [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.672667] env[62965]: DEBUG nova.compute.manager [req-4fbd61d7-8b38-459a-b821-4a733b4092aa req-28777320-e20c-4cce-9b96-11618005378e service nova] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Received event network-vif-deleted-6f417bf0-8e7f-4fa8-96c5-e21e200204c3 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 775.727194] env[62965]: INFO nova.compute.manager [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Took 26.45 seconds to build instance. [ 775.826630] env[62965]: DEBUG nova.scheduler.client.report [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 775.842407] env[62965]: DEBUG oslo_concurrency.lockutils [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Releasing lock "refresh_cache-9a2a324c-4eb1-45b6-8414-42671b412429" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.842818] env[62965]: DEBUG nova.compute.manager [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 775.843019] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 775.843571] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5d0ca968-d765-4ae5-8314-958f761ed9bc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.853065] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9b6f03f-4121-450f-bf29-df7f34310c9e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.876027] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9a2a324c-4eb1-45b6-8414-42671b412429 could not be found. [ 775.876227] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 775.876408] env[62965]: INFO nova.compute.manager [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Took 0.03 seconds to destroy the instance on the hypervisor. [ 775.876653] env[62965]: DEBUG oslo.service.loopingcall [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 775.876888] env[62965]: DEBUG nova.compute.manager [-] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 775.877061] env[62965]: DEBUG nova.network.neutron [-] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 775.893239] env[62965]: DEBUG nova.network.neutron [-] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 776.018066] env[62965]: DEBUG oslo_vmware.api [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389796, 'name': ReconfigVM_Task, 'duration_secs': 0.263222} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.018397] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Reconfigured VM instance instance-00000032 to attach disk [datastore2] e87adfae-3ea4-4cc4-9978-d8c811c80941/e87adfae-3ea4-4cc4-9978-d8c811c80941.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 776.019068] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-09dac2dd-346f-43fc-bb79-c184922c2399 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.026037] env[62965]: DEBUG oslo_vmware.api [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Waiting for the task: (returnval){ [ 776.026037] env[62965]: value = "task-1389797" [ 776.026037] env[62965]: _type = "Task" [ 776.026037] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.035054] env[62965]: DEBUG oslo_vmware.api [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389797, 'name': Rename_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.096958] env[62965]: INFO nova.compute.manager [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] [instance: cd978d56-ba80-44a9-9a02-3a1254066571] Took 1.02 seconds to deallocate network for instance. [ 776.229320] env[62965]: DEBUG oslo_concurrency.lockutils [None req-00fd8c26-55ec-40c4-b237-393fd4343a5e tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Lock "6e0a4e4b-21a2-4185-999e-69b4c74c37d7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 125.783s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.331850] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.335s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.332586] env[62965]: DEBUG nova.compute.manager [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 776.335674] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.238s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.395638] env[62965]: DEBUG nova.network.neutron [-] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.534699] env[62965]: DEBUG oslo_vmware.api [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389797, 'name': Rename_Task, 'duration_secs': 0.135685} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.534966] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 776.535246] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4542fd52-1fa9-4c2c-b9a3-76ce8f4be22f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.541483] env[62965]: DEBUG oslo_vmware.api [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Waiting for the task: (returnval){ [ 776.541483] env[62965]: value = "task-1389798" [ 776.541483] env[62965]: _type = "Task" [ 776.541483] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.548969] env[62965]: DEBUG oslo_vmware.api [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389798, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.731943] env[62965]: DEBUG nova.compute.manager [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 776.837301] env[62965]: DEBUG nova.compute.utils [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 776.838662] env[62965]: DEBUG nova.compute.manager [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 776.838836] env[62965]: DEBUG nova.network.neutron [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 776.882148] env[62965]: DEBUG nova.policy [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '34e15ad3c87445e5a217300a23223540', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8bfa5cfdc7d5496482d87675f2a122dd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 776.901354] env[62965]: INFO nova.compute.manager [-] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Took 1.02 seconds to deallocate network for instance. [ 776.904912] env[62965]: DEBUG nova.compute.claims [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 776.905113] env[62965]: DEBUG oslo_concurrency.lockutils [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.054168] env[62965]: DEBUG oslo_vmware.api [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389798, 'name': PowerOnVM_Task, 'duration_secs': 0.403939} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.054440] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 777.054639] env[62965]: INFO nova.compute.manager [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Took 5.29 seconds to spawn the instance on the hypervisor. [ 777.054813] env[62965]: DEBUG nova.compute.manager [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 777.055590] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12d1be58-7f34-40b7-b914-c10c183edc7c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.131377] env[62965]: INFO nova.scheduler.client.report [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] Deleted allocations for instance cd978d56-ba80-44a9-9a02-3a1254066571 [ 777.153793] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c647a624-062a-4f00-900b-3b6e3160277b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.162779] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f18deeb4-8469-47fa-b0a7-3c89f85ff06d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.195624] env[62965]: DEBUG nova.network.neutron [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Successfully created port: d003e564-27fb-48ac-a041-6b3c87abf7ef {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 777.198040] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06025eaa-4876-4f45-b5ae-635c370cd1fc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.206236] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb2bb37c-e760-4718-abce-cd4c744f85de {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.220080] env[62965]: DEBUG nova.compute.provider_tree [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 777.257281] env[62965]: DEBUG oslo_concurrency.lockutils [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.342266] env[62965]: DEBUG nova.compute.manager [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 777.578510] env[62965]: INFO nova.compute.manager [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Took 26.32 seconds to build instance. [ 777.642086] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f179be58-ae69-4f4b-9012-ac9addcf1697 tempest-ServerExternalEventsTest-1427201277 tempest-ServerExternalEventsTest-1427201277-project-member] Lock "cd978d56-ba80-44a9-9a02-3a1254066571" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 169.468s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.725029] env[62965]: DEBUG nova.scheduler.client.report [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 778.073008] env[62965]: DEBUG nova.compute.manager [req-aeaa9648-8362-4b86-8f07-80f462814cef req-0377d26d-582b-4d37-9df1-b85a696b3ba1 service nova] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Received event network-changed-d003e564-27fb-48ac-a041-6b3c87abf7ef {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 778.073269] env[62965]: DEBUG nova.compute.manager [req-aeaa9648-8362-4b86-8f07-80f462814cef req-0377d26d-582b-4d37-9df1-b85a696b3ba1 service nova] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Refreshing instance network info cache due to event network-changed-d003e564-27fb-48ac-a041-6b3c87abf7ef. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 778.073431] env[62965]: DEBUG oslo_concurrency.lockutils [req-aeaa9648-8362-4b86-8f07-80f462814cef req-0377d26d-582b-4d37-9df1-b85a696b3ba1 service nova] Acquiring lock "refresh_cache-dcae9cae-d4f5-4b34-b392-00b5ef9c81bf" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.073572] env[62965]: DEBUG oslo_concurrency.lockutils [req-aeaa9648-8362-4b86-8f07-80f462814cef req-0377d26d-582b-4d37-9df1-b85a696b3ba1 service nova] Acquired lock "refresh_cache-dcae9cae-d4f5-4b34-b392-00b5ef9c81bf" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.073728] env[62965]: DEBUG nova.network.neutron [req-aeaa9648-8362-4b86-8f07-80f462814cef req-0377d26d-582b-4d37-9df1-b85a696b3ba1 service nova] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Refreshing network info cache for port d003e564-27fb-48ac-a041-6b3c87abf7ef {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 778.080892] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7665ced2-42fb-4905-bb50-453dced1deef tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Lock "e87adfae-3ea4-4cc4-9978-d8c811c80941" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 127.092s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.144421] env[62965]: DEBUG nova.compute.manager [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 778.228736] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.893s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.229504] env[62965]: ERROR nova.compute.manager [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7c541fe7-1952-4a2e-8a49-bc41ca6ef389, please check neutron logs for more information. [ 778.229504] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Traceback (most recent call last): [ 778.229504] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 778.229504] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] self.driver.spawn(context, instance, image_meta, [ 778.229504] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 778.229504] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 778.229504] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 778.229504] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] vm_ref = self.build_virtual_machine(instance, [ 778.229504] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 778.229504] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] vif_infos = vmwarevif.get_vif_info(self._session, [ 778.229504] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 778.230271] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] for vif in network_info: [ 778.230271] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 778.230271] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] return self._sync_wrapper(fn, *args, **kwargs) [ 778.230271] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 778.230271] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] self.wait() [ 778.230271] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 778.230271] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] self[:] = self._gt.wait() [ 778.230271] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 778.230271] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] return self._exit_event.wait() [ 778.230271] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 778.230271] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] current.throw(*self._exc) [ 778.230271] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 778.230271] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] result = function(*args, **kwargs) [ 778.231036] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 778.231036] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] return func(*args, **kwargs) [ 778.231036] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 778.231036] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] raise e [ 778.231036] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 778.231036] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] nwinfo = self.network_api.allocate_for_instance( [ 778.231036] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 778.231036] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] created_port_ids = self._update_ports_for_instance( [ 778.231036] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 778.231036] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] with excutils.save_and_reraise_exception(): [ 778.231036] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 778.231036] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] self.force_reraise() [ 778.231036] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 778.231511] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] raise self.value [ 778.231511] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 778.231511] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] updated_port = self._update_port( [ 778.231511] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 778.231511] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] _ensure_no_port_binding_failure(port) [ 778.231511] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 778.231511] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] raise exception.PortBindingFailed(port_id=port['id']) [ 778.231511] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] nova.exception.PortBindingFailed: Binding failed for port 7c541fe7-1952-4a2e-8a49-bc41ca6ef389, please check neutron logs for more information. [ 778.231511] env[62965]: ERROR nova.compute.manager [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] [ 778.231511] env[62965]: DEBUG nova.compute.utils [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Binding failed for port 7c541fe7-1952-4a2e-8a49-bc41ca6ef389, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 778.232882] env[62965]: ERROR nova.compute.manager [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d003e564-27fb-48ac-a041-6b3c87abf7ef, please check neutron logs for more information. [ 778.232882] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 778.232882] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 778.232882] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 778.232882] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 778.232882] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 778.232882] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 778.232882] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 778.232882] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 778.232882] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 778.232882] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 778.232882] env[62965]: ERROR nova.compute.manager raise self.value [ 778.232882] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 778.232882] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 778.232882] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 778.232882] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 778.233822] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 778.233822] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 778.233822] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d003e564-27fb-48ac-a041-6b3c87abf7ef, please check neutron logs for more information. [ 778.233822] env[62965]: ERROR nova.compute.manager [ 778.233822] env[62965]: Traceback (most recent call last): [ 778.233822] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 778.233822] env[62965]: listener.cb(fileno) [ 778.233822] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 778.233822] env[62965]: result = function(*args, **kwargs) [ 778.233822] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 778.233822] env[62965]: return func(*args, **kwargs) [ 778.233822] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 778.233822] env[62965]: raise e [ 778.233822] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 778.233822] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 778.233822] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 778.233822] env[62965]: created_port_ids = self._update_ports_for_instance( [ 778.233822] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 778.233822] env[62965]: with excutils.save_and_reraise_exception(): [ 778.233822] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 778.233822] env[62965]: self.force_reraise() [ 778.233822] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 778.233822] env[62965]: raise self.value [ 778.233822] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 778.233822] env[62965]: updated_port = self._update_port( [ 778.233822] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 778.233822] env[62965]: _ensure_no_port_binding_failure(port) [ 778.233822] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 778.233822] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 778.235332] env[62965]: nova.exception.PortBindingFailed: Binding failed for port d003e564-27fb-48ac-a041-6b3c87abf7ef, please check neutron logs for more information. [ 778.235332] env[62965]: Removing descriptor: 19 [ 778.235332] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.496s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.236916] env[62965]: DEBUG nova.compute.manager [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Build of instance f9a2b957-a312-4c9b-bb0d-0b1dad6416cd was re-scheduled: Binding failed for port 7c541fe7-1952-4a2e-8a49-bc41ca6ef389, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 778.237428] env[62965]: DEBUG nova.compute.manager [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 778.237667] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Acquiring lock "refresh_cache-f9a2b957-a312-4c9b-bb0d-0b1dad6416cd" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.237814] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Acquired lock "refresh_cache-f9a2b957-a312-4c9b-bb0d-0b1dad6416cd" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.237971] env[62965]: DEBUG nova.network.neutron [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 778.352708] env[62965]: DEBUG nova.compute.manager [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 778.371168] env[62965]: INFO nova.compute.manager [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Rebuilding instance [ 778.379574] env[62965]: DEBUG nova.virt.hardware [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 778.379830] env[62965]: DEBUG nova.virt.hardware [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 778.380064] env[62965]: DEBUG nova.virt.hardware [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 778.380260] env[62965]: DEBUG nova.virt.hardware [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 778.380403] env[62965]: DEBUG nova.virt.hardware [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 778.380554] env[62965]: DEBUG nova.virt.hardware [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 778.380788] env[62965]: DEBUG nova.virt.hardware [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 778.380910] env[62965]: DEBUG nova.virt.hardware [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 778.381087] env[62965]: DEBUG nova.virt.hardware [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 778.381255] env[62965]: DEBUG nova.virt.hardware [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 778.381421] env[62965]: DEBUG nova.virt.hardware [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 778.382285] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bc8f6a2-b71b-47d7-8f05-d47c934a8657 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.394794] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-748c2774-6ee6-4ece-ba8a-eb0695e128ab {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.408545] env[62965]: ERROR nova.compute.manager [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d003e564-27fb-48ac-a041-6b3c87abf7ef, please check neutron logs for more information. [ 778.408545] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Traceback (most recent call last): [ 778.408545] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 778.408545] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] yield resources [ 778.408545] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 778.408545] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] self.driver.spawn(context, instance, image_meta, [ 778.408545] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 778.408545] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 778.408545] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 778.408545] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] vm_ref = self.build_virtual_machine(instance, [ 778.408545] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 778.409099] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] vif_infos = vmwarevif.get_vif_info(self._session, [ 778.409099] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 778.409099] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] for vif in network_info: [ 778.409099] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 778.409099] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] return self._sync_wrapper(fn, *args, **kwargs) [ 778.409099] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 778.409099] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] self.wait() [ 778.409099] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 778.409099] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] self[:] = self._gt.wait() [ 778.409099] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 778.409099] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] return self._exit_event.wait() [ 778.409099] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 778.409099] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] current.throw(*self._exc) [ 778.409873] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 778.409873] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] result = function(*args, **kwargs) [ 778.409873] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 778.409873] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] return func(*args, **kwargs) [ 778.409873] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 778.409873] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] raise e [ 778.409873] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 778.409873] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] nwinfo = self.network_api.allocate_for_instance( [ 778.409873] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 778.409873] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] created_port_ids = self._update_ports_for_instance( [ 778.409873] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 778.409873] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] with excutils.save_and_reraise_exception(): [ 778.409873] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 778.410516] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] self.force_reraise() [ 778.410516] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 778.410516] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] raise self.value [ 778.410516] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 778.410516] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] updated_port = self._update_port( [ 778.410516] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 778.410516] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] _ensure_no_port_binding_failure(port) [ 778.410516] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 778.410516] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] raise exception.PortBindingFailed(port_id=port['id']) [ 778.410516] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] nova.exception.PortBindingFailed: Binding failed for port d003e564-27fb-48ac-a041-6b3c87abf7ef, please check neutron logs for more information. [ 778.410516] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] [ 778.410516] env[62965]: INFO nova.compute.manager [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Terminating instance [ 778.417164] env[62965]: DEBUG nova.compute.manager [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 778.417923] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe557402-ad72-4e88-9208-b26ca6507686 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.583616] env[62965]: DEBUG nova.compute.manager [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 778.593237] env[62965]: DEBUG nova.network.neutron [req-aeaa9648-8362-4b86-8f07-80f462814cef req-0377d26d-582b-4d37-9df1-b85a696b3ba1 service nova] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 778.663845] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.697902] env[62965]: DEBUG nova.network.neutron [req-aeaa9648-8362-4b86-8f07-80f462814cef req-0377d26d-582b-4d37-9df1-b85a696b3ba1 service nova] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.763057] env[62965]: DEBUG nova.network.neutron [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 778.870498] env[62965]: DEBUG nova.network.neutron [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.912868] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "refresh_cache-dcae9cae-d4f5-4b34-b392-00b5ef9c81bf" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.028681] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69c700c2-da6a-4fff-b2d8-6410f62f87f0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.036736] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f712ebda-c918-4577-b2b5-e33c4a16290f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.069713] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33b84da5-21a9-4f67-8bd2-fb3d336093e6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.080325] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e56f17b-519e-4518-adfa-4900906eacf1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.097209] env[62965]: DEBUG nova.compute.provider_tree [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 779.109544] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.201098] env[62965]: DEBUG oslo_concurrency.lockutils [req-aeaa9648-8362-4b86-8f07-80f462814cef req-0377d26d-582b-4d37-9df1-b85a696b3ba1 service nova] Releasing lock "refresh_cache-dcae9cae-d4f5-4b34-b392-00b5ef9c81bf" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.201522] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquired lock "refresh_cache-dcae9cae-d4f5-4b34-b392-00b5ef9c81bf" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.201698] env[62965]: DEBUG nova.network.neutron [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 779.373336] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Releasing lock "refresh_cache-f9a2b957-a312-4c9b-bb0d-0b1dad6416cd" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.373772] env[62965]: DEBUG nova.compute.manager [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 779.374022] env[62965]: DEBUG nova.compute.manager [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 779.374343] env[62965]: DEBUG nova.network.neutron [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 779.391721] env[62965]: DEBUG nova.network.neutron [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 779.432967] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 779.433110] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ee61da59-8e4e-4a49-a6bf-6e155e585c6f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.442379] env[62965]: DEBUG oslo_vmware.api [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Waiting for the task: (returnval){ [ 779.442379] env[62965]: value = "task-1389799" [ 779.442379] env[62965]: _type = "Task" [ 779.442379] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.449383] env[62965]: DEBUG oslo_vmware.api [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389799, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.602069] env[62965]: DEBUG nova.scheduler.client.report [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 779.721179] env[62965]: DEBUG nova.network.neutron [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 779.805664] env[62965]: DEBUG nova.network.neutron [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.893818] env[62965]: DEBUG nova.network.neutron [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.951319] env[62965]: DEBUG oslo_vmware.api [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389799, 'name': PowerOffVM_Task, 'duration_secs': 0.109977} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.951584] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 779.951826] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 779.952605] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed69b28d-b084-4560-a7c2-3fdc6ad043d2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.959330] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 779.959566] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d8e8bd06-78c5-48b2-9010-cb2aee38c361 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.025849] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 780.026101] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Deleting contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 780.026287] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Deleting the datastore file [datastore2] e87adfae-3ea4-4cc4-9978-d8c811c80941 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 780.026558] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5f3166e0-791c-4b62-9d1c-dbfccd04130f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.034065] env[62965]: DEBUG oslo_vmware.api [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Waiting for the task: (returnval){ [ 780.034065] env[62965]: value = "task-1389801" [ 780.034065] env[62965]: _type = "Task" [ 780.034065] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.042209] env[62965]: DEBUG oslo_vmware.api [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389801, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.100825] env[62965]: DEBUG nova.compute.manager [req-6c71bb79-cc8a-4753-b539-1e01fc5a98b2 req-db51b47f-a677-40af-b536-03d6225bb0db service nova] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Received event network-vif-deleted-d003e564-27fb-48ac-a041-6b3c87abf7ef {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 780.107179] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.874s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.107767] env[62965]: ERROR nova.compute.manager [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c535343e-d317-4d03-ad15-9609c7687cac, please check neutron logs for more information. [ 780.107767] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Traceback (most recent call last): [ 780.107767] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 780.107767] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] self.driver.spawn(context, instance, image_meta, [ 780.107767] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 780.107767] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 780.107767] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 780.107767] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] vm_ref = self.build_virtual_machine(instance, [ 780.107767] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 780.107767] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] vif_infos = vmwarevif.get_vif_info(self._session, [ 780.107767] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 780.108194] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] for vif in network_info: [ 780.108194] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 780.108194] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] return self._sync_wrapper(fn, *args, **kwargs) [ 780.108194] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 780.108194] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] self.wait() [ 780.108194] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 780.108194] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] self[:] = self._gt.wait() [ 780.108194] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 780.108194] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] return self._exit_event.wait() [ 780.108194] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 780.108194] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] current.throw(*self._exc) [ 780.108194] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 780.108194] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] result = function(*args, **kwargs) [ 780.108640] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 780.108640] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] return func(*args, **kwargs) [ 780.108640] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 780.108640] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] raise e [ 780.108640] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 780.108640] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] nwinfo = self.network_api.allocate_for_instance( [ 780.108640] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 780.108640] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] created_port_ids = self._update_ports_for_instance( [ 780.108640] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 780.108640] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] with excutils.save_and_reraise_exception(): [ 780.108640] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 780.108640] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] self.force_reraise() [ 780.108640] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 780.109224] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] raise self.value [ 780.109224] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 780.109224] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] updated_port = self._update_port( [ 780.109224] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 780.109224] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] _ensure_no_port_binding_failure(port) [ 780.109224] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 780.109224] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] raise exception.PortBindingFailed(port_id=port['id']) [ 780.109224] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] nova.exception.PortBindingFailed: Binding failed for port c535343e-d317-4d03-ad15-9609c7687cac, please check neutron logs for more information. [ 780.109224] env[62965]: ERROR nova.compute.manager [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] [ 780.109224] env[62965]: DEBUG nova.compute.utils [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Binding failed for port c535343e-d317-4d03-ad15-9609c7687cac, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 780.109592] env[62965]: DEBUG oslo_concurrency.lockutils [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.257s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.112912] env[62965]: DEBUG nova.compute.manager [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Build of instance 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5 was re-scheduled: Binding failed for port c535343e-d317-4d03-ad15-9609c7687cac, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 780.114906] env[62965]: DEBUG nova.compute.manager [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 780.114906] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquiring lock "refresh_cache-3a994542-8c8e-4b3c-87bc-2c40a0fc95b5" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.114906] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquired lock "refresh_cache-3a994542-8c8e-4b3c-87bc-2c40a0fc95b5" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.114906] env[62965]: DEBUG nova.network.neutron [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 780.308309] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Releasing lock "refresh_cache-dcae9cae-d4f5-4b34-b392-00b5ef9c81bf" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.308732] env[62965]: DEBUG nova.compute.manager [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 780.308924] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 780.309290] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c7ecb3c2-d580-4a1b-b8aa-16cd38c857d3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.318471] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f22378ee-43d5-4d77-a3eb-eda99f550076 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.340690] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance dcae9cae-d4f5-4b34-b392-00b5ef9c81bf could not be found. [ 780.340939] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 780.341094] env[62965]: INFO nova.compute.manager [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Took 0.03 seconds to destroy the instance on the hypervisor. [ 780.341364] env[62965]: DEBUG oslo.service.loopingcall [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 780.341592] env[62965]: DEBUG nova.compute.manager [-] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 780.341690] env[62965]: DEBUG nova.network.neutron [-] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 780.362789] env[62965]: DEBUG nova.network.neutron [-] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 780.396551] env[62965]: INFO nova.compute.manager [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] [instance: f9a2b957-a312-4c9b-bb0d-0b1dad6416cd] Took 1.02 seconds to deallocate network for instance. [ 780.545313] env[62965]: DEBUG oslo_vmware.api [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389801, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.094365} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.545621] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 780.545828] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Deleted contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 780.545973] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 780.634205] env[62965]: DEBUG nova.network.neutron [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 780.725083] env[62965]: DEBUG nova.network.neutron [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.867646] env[62965]: DEBUG nova.network.neutron [-] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.879625] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2285facb-e82b-4eed-85b1-1f04fa3a90ab {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.888540] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dd8e5d9-b2cb-4191-a631-9afc4f6386cd {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.922635] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c50c6e5-68b4-4f2d-9fbe-e662594fd0ef {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.930016] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e006aea-e7bd-4015-8e31-3f84bcbe7949 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.943333] env[62965]: DEBUG nova.compute.provider_tree [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 781.228079] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Releasing lock "refresh_cache-3a994542-8c8e-4b3c-87bc-2c40a0fc95b5" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.228516] env[62965]: DEBUG nova.compute.manager [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 781.228516] env[62965]: DEBUG nova.compute.manager [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 781.228687] env[62965]: DEBUG nova.network.neutron [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 781.251091] env[62965]: DEBUG nova.network.neutron [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 781.372109] env[62965]: INFO nova.compute.manager [-] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Took 1.03 seconds to deallocate network for instance. [ 781.374397] env[62965]: DEBUG nova.compute.claims [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 781.374573] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.450266] env[62965]: DEBUG nova.scheduler.client.report [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 781.462988] env[62965]: INFO nova.scheduler.client.report [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Deleted allocations for instance f9a2b957-a312-4c9b-bb0d-0b1dad6416cd [ 781.578568] env[62965]: DEBUG nova.virt.hardware [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 781.579383] env[62965]: DEBUG nova.virt.hardware [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 781.579544] env[62965]: DEBUG nova.virt.hardware [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 781.579725] env[62965]: DEBUG nova.virt.hardware [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 781.579867] env[62965]: DEBUG nova.virt.hardware [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 781.580051] env[62965]: DEBUG nova.virt.hardware [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 781.580278] env[62965]: DEBUG nova.virt.hardware [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 781.580434] env[62965]: DEBUG nova.virt.hardware [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 781.580596] env[62965]: DEBUG nova.virt.hardware [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 781.580754] env[62965]: DEBUG nova.virt.hardware [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 781.580924] env[62965]: DEBUG nova.virt.hardware [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 781.581807] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42f2a053-7182-472c-8c2c-84fb7f1f4d40 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.589696] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a10b6728-f5de-4a40-9548-41c906a74178 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.602768] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Instance VIF info [] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 781.608101] env[62965]: DEBUG oslo.service.loopingcall [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 781.608328] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 781.608525] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a78441cc-ff07-479b-b63c-bb43b44c839e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.624862] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 781.624862] env[62965]: value = "task-1389802" [ 781.624862] env[62965]: _type = "Task" [ 781.624862] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.632411] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389802, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.753784] env[62965]: DEBUG nova.network.neutron [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.951505] env[62965]: DEBUG oslo_concurrency.lockutils [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.842s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.952218] env[62965]: ERROR nova.compute.manager [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f3d10573-dcc2-4f2b-b5bd-09bddb6d75d7, please check neutron logs for more information. [ 781.952218] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] Traceback (most recent call last): [ 781.952218] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 781.952218] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] self.driver.spawn(context, instance, image_meta, [ 781.952218] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 781.952218] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 781.952218] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 781.952218] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] vm_ref = self.build_virtual_machine(instance, [ 781.952218] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 781.952218] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] vif_infos = vmwarevif.get_vif_info(self._session, [ 781.952218] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 781.952672] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] for vif in network_info: [ 781.952672] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 781.952672] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] return self._sync_wrapper(fn, *args, **kwargs) [ 781.952672] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 781.952672] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] self.wait() [ 781.952672] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 781.952672] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] self[:] = self._gt.wait() [ 781.952672] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 781.952672] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] return self._exit_event.wait() [ 781.952672] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 781.952672] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] current.throw(*self._exc) [ 781.952672] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 781.952672] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] result = function(*args, **kwargs) [ 781.953138] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 781.953138] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] return func(*args, **kwargs) [ 781.953138] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 781.953138] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] raise e [ 781.953138] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 781.953138] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] nwinfo = self.network_api.allocate_for_instance( [ 781.953138] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 781.953138] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] created_port_ids = self._update_ports_for_instance( [ 781.953138] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 781.953138] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] with excutils.save_and_reraise_exception(): [ 781.953138] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 781.953138] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] self.force_reraise() [ 781.953138] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 781.953583] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] raise self.value [ 781.953583] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 781.953583] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] updated_port = self._update_port( [ 781.953583] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 781.953583] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] _ensure_no_port_binding_failure(port) [ 781.953583] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 781.953583] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] raise exception.PortBindingFailed(port_id=port['id']) [ 781.953583] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] nova.exception.PortBindingFailed: Binding failed for port f3d10573-dcc2-4f2b-b5bd-09bddb6d75d7, please check neutron logs for more information. [ 781.953583] env[62965]: ERROR nova.compute.manager [instance: 864ce643-8620-4405-b345-9164c81fec2a] [ 781.953892] env[62965]: DEBUG nova.compute.utils [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Binding failed for port f3d10573-dcc2-4f2b-b5bd-09bddb6d75d7, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 781.955805] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.676s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.961913] env[62965]: DEBUG nova.compute.manager [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Build of instance 864ce643-8620-4405-b345-9164c81fec2a was re-scheduled: Binding failed for port f3d10573-dcc2-4f2b-b5bd-09bddb6d75d7, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 781.961913] env[62965]: DEBUG nova.compute.manager [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 781.962106] env[62965]: DEBUG oslo_concurrency.lockutils [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Acquiring lock "refresh_cache-864ce643-8620-4405-b345-9164c81fec2a" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.962236] env[62965]: DEBUG oslo_concurrency.lockutils [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Acquired lock "refresh_cache-864ce643-8620-4405-b345-9164c81fec2a" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.962439] env[62965]: DEBUG nova.network.neutron [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 781.975511] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4e38c23d-d071-4f64-94fb-76b895cfce90 tempest-VolumesAdminNegativeTest-2050630739 tempest-VolumesAdminNegativeTest-2050630739-project-member] Lock "f9a2b957-a312-4c9b-bb0d-0b1dad6416cd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 155.969s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.136721] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389802, 'name': CreateVM_Task, 'duration_secs': 0.236687} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.137091] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 782.137662] env[62965]: DEBUG oslo_concurrency.lockutils [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.144139] env[62965]: DEBUG oslo_concurrency.lockutils [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.144139] env[62965]: DEBUG oslo_concurrency.lockutils [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 782.144139] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-caf5391b-39f0-4799-8fda-b1739b185666 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.148784] env[62965]: DEBUG oslo_vmware.api [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Waiting for the task: (returnval){ [ 782.148784] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52157d3f-bdbc-e8fc-4794-f49c93917d4c" [ 782.148784] env[62965]: _type = "Task" [ 782.148784] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.164295] env[62965]: DEBUG oslo_vmware.api [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52157d3f-bdbc-e8fc-4794-f49c93917d4c, 'name': SearchDatastore_Task, 'duration_secs': 0.00965} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.164669] env[62965]: DEBUG oslo_concurrency.lockutils [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.165122] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 782.165423] env[62965]: DEBUG oslo_concurrency.lockutils [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.165606] env[62965]: DEBUG oslo_concurrency.lockutils [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.165840] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 782.166211] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-471eb7e5-15af-45f2-b61f-3a352f7f1980 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.174198] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 782.174459] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 782.175281] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c893cf5a-b04d-46da-9a71-5d7ff418934c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.180389] env[62965]: DEBUG oslo_vmware.api [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Waiting for the task: (returnval){ [ 782.180389] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]527840eb-363a-b34d-94c9-303846f61ece" [ 782.180389] env[62965]: _type = "Task" [ 782.180389] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.188122] env[62965]: DEBUG oslo_vmware.api [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]527840eb-363a-b34d-94c9-303846f61ece, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.257948] env[62965]: INFO nova.compute.manager [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5] Took 1.03 seconds to deallocate network for instance. [ 782.478446] env[62965]: DEBUG nova.compute.manager [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 782.489748] env[62965]: DEBUG nova.network.neutron [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 782.570685] env[62965]: DEBUG nova.network.neutron [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.695811] env[62965]: DEBUG oslo_vmware.api [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]527840eb-363a-b34d-94c9-303846f61ece, 'name': SearchDatastore_Task, 'duration_secs': 0.007798} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.697160] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-79bd6e15-e1bb-4deb-997c-7abf097a1ff1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.706129] env[62965]: DEBUG oslo_vmware.api [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Waiting for the task: (returnval){ [ 782.706129] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52dea124-b118-f71a-f900-505d9027606a" [ 782.706129] env[62965]: _type = "Task" [ 782.706129] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.716021] env[62965]: DEBUG oslo_vmware.api [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52dea124-b118-f71a-f900-505d9027606a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.822376] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee6aa5bd-c574-4c6c-bab9-9d6d5a0e2228 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.831603] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84349461-b93c-4e74-9ce7-c0d9452aadf9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.862732] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6205c2db-a0e0-4725-8781-cc3cc3c44b5c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.870439] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dd8944b-91c1-45aa-9bc7-4ae659591b22 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.884204] env[62965]: DEBUG nova.compute.provider_tree [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 783.005441] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.072881] env[62965]: DEBUG oslo_concurrency.lockutils [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Releasing lock "refresh_cache-864ce643-8620-4405-b345-9164c81fec2a" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.073467] env[62965]: DEBUG nova.compute.manager [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 783.073467] env[62965]: DEBUG nova.compute.manager [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 783.073646] env[62965]: DEBUG nova.network.neutron [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: 864ce643-8620-4405-b345-9164c81fec2a] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 783.090951] env[62965]: DEBUG nova.network.neutron [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 783.219728] env[62965]: DEBUG oslo_vmware.api [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52dea124-b118-f71a-f900-505d9027606a, 'name': SearchDatastore_Task, 'duration_secs': 0.0271} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.220052] env[62965]: DEBUG oslo_concurrency.lockutils [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.220552] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] e87adfae-3ea4-4cc4-9978-d8c811c80941/e87adfae-3ea4-4cc4-9978-d8c811c80941.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 783.220669] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-28aa94fa-045e-4789-abba-37c3953db57b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.228398] env[62965]: DEBUG oslo_vmware.api [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Waiting for the task: (returnval){ [ 783.228398] env[62965]: value = "task-1389803" [ 783.228398] env[62965]: _type = "Task" [ 783.228398] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.242487] env[62965]: DEBUG oslo_vmware.api [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389803, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.307874] env[62965]: INFO nova.scheduler.client.report [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Deleted allocations for instance 3a994542-8c8e-4b3c-87bc-2c40a0fc95b5 [ 783.388956] env[62965]: DEBUG nova.scheduler.client.report [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 783.593673] env[62965]: DEBUG nova.network.neutron [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.741364] env[62965]: DEBUG oslo_vmware.api [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389803, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.480031} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.741364] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] e87adfae-3ea4-4cc4-9978-d8c811c80941/e87adfae-3ea4-4cc4-9978-d8c811c80941.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 783.741364] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 783.741364] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7d9b4059-7fd1-41de-9d5d-057d59727a7c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.749789] env[62965]: DEBUG oslo_vmware.api [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Waiting for the task: (returnval){ [ 783.749789] env[62965]: value = "task-1389804" [ 783.749789] env[62965]: _type = "Task" [ 783.749789] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.758526] env[62965]: DEBUG oslo_vmware.api [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389804, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.821580] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20ebb248-1e42-4685-9552-178b408d660b tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "3a994542-8c8e-4b3c-87bc-2c40a0fc95b5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 150.884s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.894419] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.939s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.895166] env[62965]: ERROR nova.compute.manager [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b60bf367-f412-4721-bb9f-2112fa9042de, please check neutron logs for more information. [ 783.895166] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Traceback (most recent call last): [ 783.895166] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 783.895166] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] self.driver.spawn(context, instance, image_meta, [ 783.895166] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 783.895166] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] self._vmops.spawn(context, instance, image_meta, injected_files, [ 783.895166] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 783.895166] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] vm_ref = self.build_virtual_machine(instance, [ 783.895166] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 783.895166] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] vif_infos = vmwarevif.get_vif_info(self._session, [ 783.895166] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 783.895537] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] for vif in network_info: [ 783.895537] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 783.895537] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] return self._sync_wrapper(fn, *args, **kwargs) [ 783.895537] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 783.895537] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] self.wait() [ 783.895537] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 783.895537] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] self[:] = self._gt.wait() [ 783.895537] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 783.895537] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] return self._exit_event.wait() [ 783.895537] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 783.895537] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] current.throw(*self._exc) [ 783.895537] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 783.895537] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] result = function(*args, **kwargs) [ 783.895916] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 783.895916] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] return func(*args, **kwargs) [ 783.895916] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 783.895916] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] raise e [ 783.895916] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 783.895916] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] nwinfo = self.network_api.allocate_for_instance( [ 783.895916] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 783.895916] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] created_port_ids = self._update_ports_for_instance( [ 783.895916] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 783.895916] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] with excutils.save_and_reraise_exception(): [ 783.895916] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 783.895916] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] self.force_reraise() [ 783.895916] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 783.896314] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] raise self.value [ 783.896314] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 783.896314] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] updated_port = self._update_port( [ 783.896314] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 783.896314] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] _ensure_no_port_binding_failure(port) [ 783.896314] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 783.896314] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] raise exception.PortBindingFailed(port_id=port['id']) [ 783.896314] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] nova.exception.PortBindingFailed: Binding failed for port b60bf367-f412-4721-bb9f-2112fa9042de, please check neutron logs for more information. [ 783.896314] env[62965]: ERROR nova.compute.manager [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] [ 783.896314] env[62965]: DEBUG nova.compute.utils [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Binding failed for port b60bf367-f412-4721-bb9f-2112fa9042de, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 783.900558] env[62965]: DEBUG nova.compute.manager [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Build of instance d6e90043-ebd3-4daf-acfa-545ce8462942 was re-scheduled: Binding failed for port b60bf367-f412-4721-bb9f-2112fa9042de, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 783.900558] env[62965]: DEBUG nova.compute.manager [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 783.900558] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Acquiring lock "refresh_cache-d6e90043-ebd3-4daf-acfa-545ce8462942" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.900558] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Acquired lock "refresh_cache-d6e90043-ebd3-4daf-acfa-545ce8462942" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.900922] env[62965]: DEBUG nova.network.neutron [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 783.901853] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 16.463s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.097655] env[62965]: INFO nova.compute.manager [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: 864ce643-8620-4405-b345-9164c81fec2a] Took 1.02 seconds to deallocate network for instance. [ 784.263870] env[62965]: DEBUG oslo_vmware.api [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389804, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066903} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.264147] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 784.264989] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d99092f-03ae-4cf1-a0af-a8a314fac06f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.290327] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Reconfiguring VM instance instance-00000032 to attach disk [datastore1] e87adfae-3ea4-4cc4-9978-d8c811c80941/e87adfae-3ea4-4cc4-9978-d8c811c80941.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 784.290647] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-44b72965-6a65-4d2f-be29-a9046fc3d560 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.311120] env[62965]: DEBUG oslo_vmware.api [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Waiting for the task: (returnval){ [ 784.311120] env[62965]: value = "task-1389805" [ 784.311120] env[62965]: _type = "Task" [ 784.311120] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.319015] env[62965]: DEBUG oslo_vmware.api [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389805, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.326508] env[62965]: DEBUG nova.compute.manager [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 784.436599] env[62965]: DEBUG nova.network.neutron [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 784.530177] env[62965]: DEBUG nova.network.neutron [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.820645] env[62965]: DEBUG oslo_vmware.api [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389805, 'name': ReconfigVM_Task, 'duration_secs': 0.263602} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.820937] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Reconfigured VM instance instance-00000032 to attach disk [datastore1] e87adfae-3ea4-4cc4-9978-d8c811c80941/e87adfae-3ea4-4cc4-9978-d8c811c80941.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 784.821570] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a3c25627-f733-4f86-817f-3a8aa6f6d7ab {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.828017] env[62965]: DEBUG oslo_vmware.api [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Waiting for the task: (returnval){ [ 784.828017] env[62965]: value = "task-1389806" [ 784.828017] env[62965]: _type = "Task" [ 784.828017] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.839150] env[62965]: DEBUG oslo_vmware.api [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389806, 'name': Rename_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.856965] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.014074] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Releasing lock "refresh_cache-d6e90043-ebd3-4daf-acfa-545ce8462942" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.014074] env[62965]: DEBUG nova.compute.manager [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 785.014074] env[62965]: DEBUG nova.compute.manager [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 785.014074] env[62965]: DEBUG nova.network.neutron [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 785.032667] env[62965]: DEBUG nova.network.neutron [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 785.113648] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "c2448b76-1553-4b68-a731-0bb0967f4c1d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.113902] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "c2448b76-1553-4b68-a731-0bb0967f4c1d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.134492] env[62965]: INFO nova.scheduler.client.report [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Deleted allocations for instance 864ce643-8620-4405-b345-9164c81fec2a [ 785.341808] env[62965]: DEBUG oslo_vmware.api [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389806, 'name': Rename_Task, 'duration_secs': 0.135467} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.342126] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 785.342388] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-579be25a-3f64-4e8d-afec-7cea9372fb73 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.350384] env[62965]: DEBUG oslo_vmware.api [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Waiting for the task: (returnval){ [ 785.350384] env[62965]: value = "task-1389807" [ 785.350384] env[62965]: _type = "Task" [ 785.350384] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.358322] env[62965]: DEBUG oslo_vmware.api [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389807, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.442417] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 864ce643-8620-4405-b345-9164c81fec2a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 785.535592] env[62965]: DEBUG nova.network.neutron [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.643779] env[62965]: DEBUG oslo_concurrency.lockutils [None req-41dc0875-322c-4c79-ac33-039a17589932 tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Lock "864ce643-8620-4405-b345-9164c81fec2a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 148.296s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.647500] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquiring lock "d4d9b202-3aa7-4b66-949d-4a2eb34b3adf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.647723] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "d4d9b202-3aa7-4b66-949d-4a2eb34b3adf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.859884] env[62965]: DEBUG oslo_vmware.api [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389807, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.946703] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance d6e90043-ebd3-4daf-acfa-545ce8462942 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 785.946956] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance afaec369-6aef-4bdc-b886-eed795644867 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 785.947118] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 6e0a4e4b-21a2-4185-999e-69b4c74c37d7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 785.947247] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance e87adfae-3ea4-4cc4-9978-d8c811c80941 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 785.947364] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 9a2a324c-4eb1-45b6-8414-42671b412429 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 785.947588] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance dcae9cae-d4f5-4b34-b392-00b5ef9c81bf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 786.038406] env[62965]: INFO nova.compute.manager [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] [instance: d6e90043-ebd3-4daf-acfa-545ce8462942] Took 1.03 seconds to deallocate network for instance. [ 786.149458] env[62965]: DEBUG nova.compute.manager [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 786.360843] env[62965]: DEBUG oslo_vmware.api [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389807, 'name': PowerOnVM_Task, 'duration_secs': 0.527257} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.361139] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 786.361343] env[62965]: DEBUG nova.compute.manager [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 786.362141] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-730a0174-62dd-4295-a1ad-7ec494ee6fbe {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.450866] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 786.683227] env[62965]: DEBUG oslo_concurrency.lockutils [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.883970] env[62965]: DEBUG oslo_concurrency.lockutils [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.955107] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance ab4da02e-f3ec-473b-968f-14b102780ddb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 787.069016] env[62965]: INFO nova.scheduler.client.report [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Deleted allocations for instance d6e90043-ebd3-4daf-acfa-545ce8462942 [ 787.458056] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance c8402430-4229-4b61-9118-ddeab4d8cd79 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 787.496715] env[62965]: DEBUG oslo_concurrency.lockutils [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Acquiring lock "e87adfae-3ea4-4cc4-9978-d8c811c80941" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.496959] env[62965]: DEBUG oslo_concurrency.lockutils [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Lock "e87adfae-3ea4-4cc4-9978-d8c811c80941" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.497215] env[62965]: DEBUG oslo_concurrency.lockutils [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Acquiring lock "e87adfae-3ea4-4cc4-9978-d8c811c80941-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.497404] env[62965]: DEBUG oslo_concurrency.lockutils [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Lock "e87adfae-3ea4-4cc4-9978-d8c811c80941-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.497569] env[62965]: DEBUG oslo_concurrency.lockutils [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Lock "e87adfae-3ea4-4cc4-9978-d8c811c80941-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.499647] env[62965]: INFO nova.compute.manager [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Terminating instance [ 787.575415] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f653a15f-5633-42a6-91a4-a18df25d142e tempest-ListImageFiltersTestJSON-700321432 tempest-ListImageFiltersTestJSON-700321432-project-member] Lock "d6e90043-ebd3-4daf-acfa-545ce8462942" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 149.985s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.963878] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 788.006062] env[62965]: DEBUG oslo_concurrency.lockutils [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Acquiring lock "refresh_cache-e87adfae-3ea4-4cc4-9978-d8c811c80941" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.006062] env[62965]: DEBUG oslo_concurrency.lockutils [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Acquired lock "refresh_cache-e87adfae-3ea4-4cc4-9978-d8c811c80941" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.006062] env[62965]: DEBUG nova.network.neutron [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 788.078235] env[62965]: DEBUG nova.compute.manager [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 788.467234] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 21bf78fb-4da7-4484-9678-b22a6d944c0b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 788.524767] env[62965]: DEBUG nova.network.neutron [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 788.586019] env[62965]: DEBUG nova.network.neutron [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.611780] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.971470] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 13d20f82-fa4f-41ab-b08f-5b64da7dddc0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 789.093332] env[62965]: DEBUG oslo_concurrency.lockutils [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Releasing lock "refresh_cache-e87adfae-3ea4-4cc4-9978-d8c811c80941" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.093811] env[62965]: DEBUG nova.compute.manager [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 789.094011] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 789.094931] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86536839-7386-4e91-b7a9-2eda454eb238 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.103831] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 789.104135] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6a07d811-2116-4ce8-b404-cf210574c6d2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.113021] env[62965]: DEBUG oslo_vmware.api [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Waiting for the task: (returnval){ [ 789.113021] env[62965]: value = "task-1389808" [ 789.113021] env[62965]: _type = "Task" [ 789.113021] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.120425] env[62965]: DEBUG oslo_vmware.api [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389808, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.475615] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 474c8faf-cf69-4296-8fba-521e5472c071 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 789.622291] env[62965]: DEBUG oslo_vmware.api [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389808, 'name': PowerOffVM_Task, 'duration_secs': 0.201545} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.622556] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 789.622720] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 789.622960] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-36147104-a5f6-4462-91db-8e831f044267 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.648248] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 789.648248] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Deleting contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 789.648248] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Deleting the datastore file [datastore1] e87adfae-3ea4-4cc4-9978-d8c811c80941 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 789.648248] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a54a8d3f-888a-4342-93d6-8a5a5b69fdf6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.653370] env[62965]: DEBUG oslo_vmware.api [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Waiting for the task: (returnval){ [ 789.653370] env[62965]: value = "task-1389810" [ 789.653370] env[62965]: _type = "Task" [ 789.653370] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.661454] env[62965]: DEBUG oslo_vmware.api [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389810, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.980278] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance e9cac9cf-c165-428b-9d83-e37905f42c93 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 790.163294] env[62965]: DEBUG oslo_vmware.api [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389810, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.084718} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.163383] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 790.163560] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Deleted contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 790.164322] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 790.164322] env[62965]: INFO nova.compute.manager [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Took 1.07 seconds to destroy the instance on the hypervisor. [ 790.164322] env[62965]: DEBUG oslo.service.loopingcall [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 790.164489] env[62965]: DEBUG nova.compute.manager [-] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 790.164584] env[62965]: DEBUG nova.network.neutron [-] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 790.181636] env[62965]: DEBUG nova.network.neutron [-] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 790.483935] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 3d0f74d4-ff59-462c-9577-c01c4f15ea1d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 790.685373] env[62965]: DEBUG nova.network.neutron [-] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.987967] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 9566eb77-5198-47b3-a044-d2ed34e00888 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 791.152452] env[62965]: DEBUG oslo_concurrency.lockutils [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Acquiring lock "b0d8f38a-86d3-466a-9b83-35145071dc21" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.152690] env[62965]: DEBUG oslo_concurrency.lockutils [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Lock "b0d8f38a-86d3-466a-9b83-35145071dc21" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.187724] env[62965]: INFO nova.compute.manager [-] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Took 1.02 seconds to deallocate network for instance. [ 791.490801] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 2b6efa32-fc5b-402c-a551-b6e5c01eca2c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 791.697759] env[62965]: DEBUG oslo_concurrency.lockutils [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.993748] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 64cf898c-2c97-4daa-80c8-0a1c252cbcf2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 792.496523] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 792.999973] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance c2448b76-1553-4b68-a731-0bb0967f4c1d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 793.000405] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=62965) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 793.000583] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=62965) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 793.223983] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d83f0f17-2926-4073-9172-f97028f99e2b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.231355] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb6f616c-f30c-4a28-86e3-0b6ca3dbd95a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.261081] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9495aae6-a313-46a9-8f69-97ebb69ac2d5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.267896] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-568871e8-c788-43f5-86a5-08d4c91d298a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.280600] env[62965]: DEBUG nova.compute.provider_tree [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 793.783334] env[62965]: DEBUG nova.scheduler.client.report [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 794.288380] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62965) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 794.288627] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 10.388s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.288898] env[62965]: DEBUG oslo_concurrency.lockutils [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.620s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.291840] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 794.291987] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Cleaning up deleted instances {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11557}} [ 794.797993] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] There are 2 instances to clean {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11566}} [ 794.798300] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 02f113d2-f991-4268-b9a4-59e2e6bcf7a6] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 795.142649] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89c83e1f-7dc0-4347-afba-9d32c572a172 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.149889] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d35680af-3e6a-4a06-9fce-bc9100e43cfa {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.179657] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c8f91e4-17b8-4bce-9a7a-1641e8f31816 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.187047] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddccf5c5-77be-424b-aee7-0932c75db1d8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.199864] env[62965]: DEBUG nova.compute.provider_tree [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 795.301821] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 532ae225-da05-4cbb-9bb4-c72a8451acff] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 795.703758] env[62965]: DEBUG nova.scheduler.client.report [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 795.805138] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 795.805450] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Cleaning up deleted instances with incomplete migration {{(pid=62965) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11595}} [ 796.208633] env[62965]: DEBUG oslo_concurrency.lockutils [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.920s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.209425] env[62965]: ERROR nova.compute.manager [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: afaec369-6aef-4bdc-b886-eed795644867] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 316fb176-0010-4e0d-8df1-d84208f52dab, please check neutron logs for more information. [ 796.209425] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] Traceback (most recent call last): [ 796.209425] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 796.209425] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] self.driver.spawn(context, instance, image_meta, [ 796.209425] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 796.209425] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] self._vmops.spawn(context, instance, image_meta, injected_files, [ 796.209425] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 796.209425] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] vm_ref = self.build_virtual_machine(instance, [ 796.209425] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 796.209425] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] vif_infos = vmwarevif.get_vif_info(self._session, [ 796.209425] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 796.209752] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] for vif in network_info: [ 796.209752] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 796.209752] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] return self._sync_wrapper(fn, *args, **kwargs) [ 796.209752] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 796.209752] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] self.wait() [ 796.209752] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 796.209752] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] self[:] = self._gt.wait() [ 796.209752] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 796.209752] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] return self._exit_event.wait() [ 796.209752] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 796.209752] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] current.throw(*self._exc) [ 796.209752] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 796.209752] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] result = function(*args, **kwargs) [ 796.210082] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 796.210082] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] return func(*args, **kwargs) [ 796.210082] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 796.210082] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] raise e [ 796.210082] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 796.210082] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] nwinfo = self.network_api.allocate_for_instance( [ 796.210082] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 796.210082] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] created_port_ids = self._update_ports_for_instance( [ 796.210082] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 796.210082] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] with excutils.save_and_reraise_exception(): [ 796.210082] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 796.210082] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] self.force_reraise() [ 796.210082] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 796.210415] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] raise self.value [ 796.210415] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 796.210415] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] updated_port = self._update_port( [ 796.210415] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 796.210415] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] _ensure_no_port_binding_failure(port) [ 796.210415] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 796.210415] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] raise exception.PortBindingFailed(port_id=port['id']) [ 796.210415] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] nova.exception.PortBindingFailed: Binding failed for port 316fb176-0010-4e0d-8df1-d84208f52dab, please check neutron logs for more information. [ 796.210415] env[62965]: ERROR nova.compute.manager [instance: afaec369-6aef-4bdc-b886-eed795644867] [ 796.210415] env[62965]: DEBUG nova.compute.utils [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: afaec369-6aef-4bdc-b886-eed795644867] Binding failed for port 316fb176-0010-4e0d-8df1-d84208f52dab, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 796.211506] env[62965]: DEBUG oslo_concurrency.lockutils [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.306s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.215390] env[62965]: DEBUG nova.compute.manager [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: afaec369-6aef-4bdc-b886-eed795644867] Build of instance afaec369-6aef-4bdc-b886-eed795644867 was re-scheduled: Binding failed for port 316fb176-0010-4e0d-8df1-d84208f52dab, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 796.215813] env[62965]: DEBUG nova.compute.manager [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: afaec369-6aef-4bdc-b886-eed795644867] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 796.216050] env[62965]: DEBUG oslo_concurrency.lockutils [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Acquiring lock "refresh_cache-afaec369-6aef-4bdc-b886-eed795644867" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.216202] env[62965]: DEBUG oslo_concurrency.lockutils [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Acquired lock "refresh_cache-afaec369-6aef-4bdc-b886-eed795644867" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.216359] env[62965]: DEBUG nova.network.neutron [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: afaec369-6aef-4bdc-b886-eed795644867] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 796.307592] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 796.733160] env[62965]: DEBUG nova.scheduler.client.report [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Refreshing inventories for resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 796.735809] env[62965]: DEBUG nova.network.neutron [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: afaec369-6aef-4bdc-b886-eed795644867] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 796.748143] env[62965]: DEBUG nova.scheduler.client.report [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Updating ProviderTree inventory for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 796.748407] env[62965]: DEBUG nova.compute.provider_tree [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 796.759119] env[62965]: DEBUG nova.scheduler.client.report [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Refreshing aggregate associations for resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8, aggregates: None {{(pid=62965) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 796.776009] env[62965]: DEBUG nova.scheduler.client.report [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Refreshing trait associations for resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64 {{(pid=62965) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 796.810669] env[62965]: DEBUG nova.network.neutron [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: afaec369-6aef-4bdc-b886-eed795644867] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.001762] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bed415a-ab48-4719-8c54-7f4c365f6c43 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.009336] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f4229e5-c315-4f21-85f4-cff8b83d06e8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.038018] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c78189ad-b2ad-479b-985f-964bffd08a26 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.044571] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e700f10-a5af-465b-95be-4fe8dba6d6a6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.056986] env[62965]: DEBUG nova.compute.provider_tree [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 797.314477] env[62965]: DEBUG oslo_concurrency.lockutils [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Releasing lock "refresh_cache-afaec369-6aef-4bdc-b886-eed795644867" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.314793] env[62965]: DEBUG nova.compute.manager [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 797.315028] env[62965]: DEBUG nova.compute.manager [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: afaec369-6aef-4bdc-b886-eed795644867] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 797.315203] env[62965]: DEBUG nova.network.neutron [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: afaec369-6aef-4bdc-b886-eed795644867] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 797.329681] env[62965]: DEBUG nova.network.neutron [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: afaec369-6aef-4bdc-b886-eed795644867] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 797.560290] env[62965]: DEBUG nova.scheduler.client.report [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 797.833349] env[62965]: DEBUG nova.network.neutron [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: afaec369-6aef-4bdc-b886-eed795644867] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.065209] env[62965]: DEBUG oslo_concurrency.lockutils [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.854s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.065856] env[62965]: ERROR nova.compute.manager [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6f417bf0-8e7f-4fa8-96c5-e21e200204c3, please check neutron logs for more information. [ 798.065856] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Traceback (most recent call last): [ 798.065856] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 798.065856] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] self.driver.spawn(context, instance, image_meta, [ 798.065856] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 798.065856] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] self._vmops.spawn(context, instance, image_meta, injected_files, [ 798.065856] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 798.065856] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] vm_ref = self.build_virtual_machine(instance, [ 798.065856] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 798.065856] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] vif_infos = vmwarevif.get_vif_info(self._session, [ 798.065856] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 798.066164] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] for vif in network_info: [ 798.066164] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 798.066164] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] return self._sync_wrapper(fn, *args, **kwargs) [ 798.066164] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 798.066164] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] self.wait() [ 798.066164] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 798.066164] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] self[:] = self._gt.wait() [ 798.066164] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 798.066164] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] return self._exit_event.wait() [ 798.066164] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 798.066164] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] current.throw(*self._exc) [ 798.066164] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 798.066164] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] result = function(*args, **kwargs) [ 798.066486] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 798.066486] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] return func(*args, **kwargs) [ 798.066486] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 798.066486] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] raise e [ 798.066486] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 798.066486] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] nwinfo = self.network_api.allocate_for_instance( [ 798.066486] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 798.066486] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] created_port_ids = self._update_ports_for_instance( [ 798.066486] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 798.066486] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] with excutils.save_and_reraise_exception(): [ 798.066486] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 798.066486] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] self.force_reraise() [ 798.066486] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 798.066993] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] raise self.value [ 798.066993] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 798.066993] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] updated_port = self._update_port( [ 798.066993] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 798.066993] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] _ensure_no_port_binding_failure(port) [ 798.066993] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 798.066993] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] raise exception.PortBindingFailed(port_id=port['id']) [ 798.066993] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] nova.exception.PortBindingFailed: Binding failed for port 6f417bf0-8e7f-4fa8-96c5-e21e200204c3, please check neutron logs for more information. [ 798.066993] env[62965]: ERROR nova.compute.manager [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] [ 798.066993] env[62965]: DEBUG nova.compute.utils [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Binding failed for port 6f417bf0-8e7f-4fa8-96c5-e21e200204c3, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 798.067898] env[62965]: DEBUG oslo_concurrency.lockutils [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.811s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.069443] env[62965]: INFO nova.compute.claims [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 798.072377] env[62965]: DEBUG nova.compute.manager [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Build of instance 9a2a324c-4eb1-45b6-8414-42671b412429 was re-scheduled: Binding failed for port 6f417bf0-8e7f-4fa8-96c5-e21e200204c3, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 798.072480] env[62965]: DEBUG nova.compute.manager [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 798.072680] env[62965]: DEBUG oslo_concurrency.lockutils [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Acquiring lock "refresh_cache-9a2a324c-4eb1-45b6-8414-42671b412429" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.072824] env[62965]: DEBUG oslo_concurrency.lockutils [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Acquired lock "refresh_cache-9a2a324c-4eb1-45b6-8414-42671b412429" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.072976] env[62965]: DEBUG nova.network.neutron [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 798.335847] env[62965]: INFO nova.compute.manager [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] [instance: afaec369-6aef-4bdc-b886-eed795644867] Took 1.02 seconds to deallocate network for instance. [ 798.590646] env[62965]: DEBUG nova.network.neutron [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 798.658490] env[62965]: DEBUG nova.network.neutron [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.160951] env[62965]: DEBUG oslo_concurrency.lockutils [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Releasing lock "refresh_cache-9a2a324c-4eb1-45b6-8414-42671b412429" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.161204] env[62965]: DEBUG nova.compute.manager [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 799.161432] env[62965]: DEBUG nova.compute.manager [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 799.161650] env[62965]: DEBUG nova.network.neutron [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 799.177292] env[62965]: DEBUG nova.network.neutron [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 799.317752] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb38981d-6878-4f24-bf68-fa4b878542d5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.325368] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-282a4884-5ca9-4f53-a0b6-027c04320dc3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.358100] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a504d34-7ecb-426f-b3a4-bbd6eba48d15 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.364899] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dff7e57-c393-440b-866a-0a1f64b605ba {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.377894] env[62965]: DEBUG nova.compute.provider_tree [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 799.379764] env[62965]: INFO nova.scheduler.client.report [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Deleted allocations for instance afaec369-6aef-4bdc-b886-eed795644867 [ 799.680524] env[62965]: DEBUG nova.network.neutron [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.886509] env[62965]: DEBUG nova.scheduler.client.report [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 799.890136] env[62965]: DEBUG oslo_concurrency.lockutils [None req-246d5221-7f9e-409a-acdf-f1cbd86ffcd0 tempest-SecurityGroupsTestJSON-1984825249 tempest-SecurityGroupsTestJSON-1984825249-project-member] Lock "afaec369-6aef-4bdc-b886-eed795644867" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 159.707s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.183065] env[62965]: INFO nova.compute.manager [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] [instance: 9a2a324c-4eb1-45b6-8414-42671b412429] Took 1.02 seconds to deallocate network for instance. [ 800.392151] env[62965]: DEBUG oslo_concurrency.lockutils [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.324s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.392644] env[62965]: DEBUG nova.compute.manager [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 800.395193] env[62965]: DEBUG nova.compute.manager [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 800.397652] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.734s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.399010] env[62965]: INFO nova.compute.claims [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 800.905627] env[62965]: DEBUG nova.compute.utils [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 800.910615] env[62965]: DEBUG nova.compute.manager [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 800.910742] env[62965]: DEBUG nova.network.neutron [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 800.925489] env[62965]: DEBUG oslo_concurrency.lockutils [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.957008] env[62965]: DEBUG nova.policy [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '64ab301ec51d40788a6ed6d0876ce256', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cce6c4888d674aa7af60cb89e6a2567a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 801.206980] env[62965]: DEBUG nova.network.neutron [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Successfully created port: ac88ed76-44e1-46ea-9dc4-31e693ea44a6 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 801.212642] env[62965]: INFO nova.scheduler.client.report [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Deleted allocations for instance 9a2a324c-4eb1-45b6-8414-42671b412429 [ 801.411782] env[62965]: DEBUG nova.compute.manager [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 801.668287] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ffb719e-8956-444d-976a-aad13168b111 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.676478] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e35c9ffc-9255-418e-8523-32349b9511a3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.711846] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3536839-1b15-49f6-8a1c-7c78e38b4d08 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.721374] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4012bd81-b9bb-40a5-bc60-c511c8dcbbfc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.725326] env[62965]: DEBUG oslo_concurrency.lockutils [None req-83216973-f5d9-4c15-84cc-dc3fb143c63c tempest-ServerMetadataNegativeTestJSON-552614043 tempest-ServerMetadataNegativeTestJSON-552614043-project-member] Lock "9a2a324c-4eb1-45b6-8414-42671b412429" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 149.813s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.738100] env[62965]: DEBUG nova.compute.provider_tree [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 802.228310] env[62965]: DEBUG nova.compute.manager [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 802.240636] env[62965]: DEBUG nova.scheduler.client.report [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 802.309113] env[62965]: DEBUG nova.compute.manager [req-f4e467b2-9723-480c-95c8-c4f4b091959e req-0794164f-d1e5-4d8c-a130-fb373c0f851d service nova] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Received event network-changed-ac88ed76-44e1-46ea-9dc4-31e693ea44a6 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 802.309113] env[62965]: DEBUG nova.compute.manager [req-f4e467b2-9723-480c-95c8-c4f4b091959e req-0794164f-d1e5-4d8c-a130-fb373c0f851d service nova] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Refreshing instance network info cache due to event network-changed-ac88ed76-44e1-46ea-9dc4-31e693ea44a6. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 802.309113] env[62965]: DEBUG oslo_concurrency.lockutils [req-f4e467b2-9723-480c-95c8-c4f4b091959e req-0794164f-d1e5-4d8c-a130-fb373c0f851d service nova] Acquiring lock "refresh_cache-6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.309113] env[62965]: DEBUG oslo_concurrency.lockutils [req-f4e467b2-9723-480c-95c8-c4f4b091959e req-0794164f-d1e5-4d8c-a130-fb373c0f851d service nova] Acquired lock "refresh_cache-6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.309113] env[62965]: DEBUG nova.network.neutron [req-f4e467b2-9723-480c-95c8-c4f4b091959e req-0794164f-d1e5-4d8c-a130-fb373c0f851d service nova] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Refreshing network info cache for port ac88ed76-44e1-46ea-9dc4-31e693ea44a6 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 802.426211] env[62965]: DEBUG nova.compute.manager [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 802.453378] env[62965]: DEBUG nova.virt.hardware [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 802.454016] env[62965]: DEBUG nova.virt.hardware [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 802.454210] env[62965]: DEBUG nova.virt.hardware [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 802.457638] env[62965]: DEBUG nova.virt.hardware [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 802.457638] env[62965]: DEBUG nova.virt.hardware [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 802.457638] env[62965]: DEBUG nova.virt.hardware [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 802.457638] env[62965]: DEBUG nova.virt.hardware [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 802.457638] env[62965]: DEBUG nova.virt.hardware [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 802.458625] env[62965]: DEBUG nova.virt.hardware [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 802.458625] env[62965]: DEBUG nova.virt.hardware [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 802.458625] env[62965]: DEBUG nova.virt.hardware [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 802.458625] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8c79f43-ec9a-4993-876a-084b0aa19737 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.467845] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24eeac0b-a94e-4170-a9ca-e16210fd646f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.696008] env[62965]: ERROR nova.compute.manager [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ac88ed76-44e1-46ea-9dc4-31e693ea44a6, please check neutron logs for more information. [ 802.696008] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 802.696008] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 802.696008] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 802.696008] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 802.696008] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 802.696008] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 802.696008] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 802.696008] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 802.696008] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 802.696008] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 802.696008] env[62965]: ERROR nova.compute.manager raise self.value [ 802.696008] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 802.696008] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 802.696008] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 802.696008] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 802.696456] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 802.696456] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 802.696456] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ac88ed76-44e1-46ea-9dc4-31e693ea44a6, please check neutron logs for more information. [ 802.696456] env[62965]: ERROR nova.compute.manager [ 802.696456] env[62965]: Traceback (most recent call last): [ 802.696456] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 802.696456] env[62965]: listener.cb(fileno) [ 802.696456] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 802.696456] env[62965]: result = function(*args, **kwargs) [ 802.696456] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 802.696456] env[62965]: return func(*args, **kwargs) [ 802.696456] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 802.696456] env[62965]: raise e [ 802.696456] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 802.696456] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 802.696456] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 802.696456] env[62965]: created_port_ids = self._update_ports_for_instance( [ 802.696456] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 802.696456] env[62965]: with excutils.save_and_reraise_exception(): [ 802.696456] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 802.696456] env[62965]: self.force_reraise() [ 802.696456] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 802.696456] env[62965]: raise self.value [ 802.696456] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 802.696456] env[62965]: updated_port = self._update_port( [ 802.696456] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 802.696456] env[62965]: _ensure_no_port_binding_failure(port) [ 802.696456] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 802.696456] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 802.697208] env[62965]: nova.exception.PortBindingFailed: Binding failed for port ac88ed76-44e1-46ea-9dc4-31e693ea44a6, please check neutron logs for more information. [ 802.697208] env[62965]: Removing descriptor: 19 [ 802.697208] env[62965]: ERROR nova.compute.manager [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ac88ed76-44e1-46ea-9dc4-31e693ea44a6, please check neutron logs for more information. [ 802.697208] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Traceback (most recent call last): [ 802.697208] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 802.697208] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] yield resources [ 802.697208] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 802.697208] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] self.driver.spawn(context, instance, image_meta, [ 802.697208] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 802.697208] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 802.697208] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 802.697208] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] vm_ref = self.build_virtual_machine(instance, [ 802.697511] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 802.697511] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] vif_infos = vmwarevif.get_vif_info(self._session, [ 802.697511] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 802.697511] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] for vif in network_info: [ 802.697511] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 802.697511] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] return self._sync_wrapper(fn, *args, **kwargs) [ 802.697511] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 802.697511] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] self.wait() [ 802.697511] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 802.697511] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] self[:] = self._gt.wait() [ 802.697511] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 802.697511] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] return self._exit_event.wait() [ 802.697511] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 802.697830] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] result = hub.switch() [ 802.697830] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 802.697830] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] return self.greenlet.switch() [ 802.697830] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 802.697830] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] result = function(*args, **kwargs) [ 802.697830] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 802.697830] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] return func(*args, **kwargs) [ 802.697830] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 802.697830] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] raise e [ 802.697830] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 802.697830] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] nwinfo = self.network_api.allocate_for_instance( [ 802.697830] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 802.697830] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] created_port_ids = self._update_ports_for_instance( [ 802.698161] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 802.698161] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] with excutils.save_and_reraise_exception(): [ 802.698161] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 802.698161] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] self.force_reraise() [ 802.698161] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 802.698161] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] raise self.value [ 802.698161] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 802.698161] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] updated_port = self._update_port( [ 802.698161] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 802.698161] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] _ensure_no_port_binding_failure(port) [ 802.698161] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 802.698161] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] raise exception.PortBindingFailed(port_id=port['id']) [ 802.698519] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] nova.exception.PortBindingFailed: Binding failed for port ac88ed76-44e1-46ea-9dc4-31e693ea44a6, please check neutron logs for more information. [ 802.698519] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] [ 802.698519] env[62965]: INFO nova.compute.manager [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Terminating instance [ 802.747409] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.350s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.748149] env[62965]: DEBUG nova.compute.manager [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 802.752151] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.643s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.755620] env[62965]: INFO nova.compute.claims [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 802.760609] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.838846] env[62965]: DEBUG nova.network.neutron [req-f4e467b2-9723-480c-95c8-c4f4b091959e req-0794164f-d1e5-4d8c-a130-fb373c0f851d service nova] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 802.970307] env[62965]: DEBUG nova.network.neutron [req-f4e467b2-9723-480c-95c8-c4f4b091959e req-0794164f-d1e5-4d8c-a130-fb373c0f851d service nova] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.203030] env[62965]: DEBUG oslo_concurrency.lockutils [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] Acquiring lock "refresh_cache-6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.265198] env[62965]: DEBUG nova.compute.utils [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 803.269412] env[62965]: DEBUG nova.compute.manager [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 803.271265] env[62965]: DEBUG nova.network.neutron [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 803.339519] env[62965]: DEBUG nova.policy [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '785dd107c5ac4f6290f2a1c241f9bb34', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ae943aa0186c4c369a2044f6c00a8fe6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 803.474765] env[62965]: DEBUG oslo_concurrency.lockutils [req-f4e467b2-9723-480c-95c8-c4f4b091959e req-0794164f-d1e5-4d8c-a130-fb373c0f851d service nova] Releasing lock "refresh_cache-6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.474765] env[62965]: DEBUG oslo_concurrency.lockutils [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] Acquired lock "refresh_cache-6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.474765] env[62965]: DEBUG nova.network.neutron [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 803.710982] env[62965]: DEBUG nova.network.neutron [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Successfully created port: 71c05380-b654-4e8a-addc-e438efd4e9ca {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 803.780286] env[62965]: DEBUG nova.compute.manager [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 803.992866] env[62965]: DEBUG nova.network.neutron [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 804.086886] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee634344-3f0b-4f22-ba09-b1b953441b5f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.094827] env[62965]: DEBUG nova.network.neutron [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.096634] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6a71eb4-d16a-4273-b824-4806fa8ac64d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.129578] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-296396b3-07f7-4316-9bb3-6d8e5d3fde9e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.137590] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-549abe9b-753b-406b-adda-2bb057ffb93d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.152245] env[62965]: DEBUG nova.compute.provider_tree [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 804.381359] env[62965]: DEBUG nova.compute.manager [req-94aedde1-5383-4d13-b609-fc6ea0c47934 req-18884076-8030-430b-b5a7-6138d167bbd3 service nova] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Received event network-vif-deleted-ac88ed76-44e1-46ea-9dc4-31e693ea44a6 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 804.601255] env[62965]: DEBUG oslo_concurrency.lockutils [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] Releasing lock "refresh_cache-6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.601644] env[62965]: DEBUG nova.compute.manager [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 804.601831] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 804.602145] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aaf65a18-8f96-44de-8178-b13378940762 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.613478] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a0ccdb6-d90f-429e-9a51-162e841a5793 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.646100] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1 could not be found. [ 804.646278] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 804.646461] env[62965]: INFO nova.compute.manager [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Took 0.04 seconds to destroy the instance on the hypervisor. [ 804.646708] env[62965]: DEBUG oslo.service.loopingcall [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 804.646931] env[62965]: DEBUG nova.compute.manager [-] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 804.647114] env[62965]: DEBUG nova.network.neutron [-] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 804.655609] env[62965]: DEBUG nova.scheduler.client.report [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 804.674391] env[62965]: DEBUG nova.network.neutron [-] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 804.688476] env[62965]: ERROR nova.compute.manager [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 71c05380-b654-4e8a-addc-e438efd4e9ca, please check neutron logs for more information. [ 804.688476] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 804.688476] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 804.688476] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 804.688476] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 804.688476] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 804.688476] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 804.688476] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 804.688476] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 804.688476] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 804.688476] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 804.688476] env[62965]: ERROR nova.compute.manager raise self.value [ 804.688476] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 804.688476] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 804.688476] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 804.688476] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 804.691749] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 804.691749] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 804.691749] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 71c05380-b654-4e8a-addc-e438efd4e9ca, please check neutron logs for more information. [ 804.691749] env[62965]: ERROR nova.compute.manager [ 804.691749] env[62965]: Traceback (most recent call last): [ 804.691749] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 804.691749] env[62965]: listener.cb(fileno) [ 804.691749] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 804.691749] env[62965]: result = function(*args, **kwargs) [ 804.691749] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 804.691749] env[62965]: return func(*args, **kwargs) [ 804.691749] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 804.691749] env[62965]: raise e [ 804.691749] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 804.691749] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 804.691749] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 804.691749] env[62965]: created_port_ids = self._update_ports_for_instance( [ 804.691749] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 804.691749] env[62965]: with excutils.save_and_reraise_exception(): [ 804.691749] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 804.691749] env[62965]: self.force_reraise() [ 804.691749] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 804.691749] env[62965]: raise self.value [ 804.691749] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 804.691749] env[62965]: updated_port = self._update_port( [ 804.691749] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 804.691749] env[62965]: _ensure_no_port_binding_failure(port) [ 804.691749] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 804.691749] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 804.692486] env[62965]: nova.exception.PortBindingFailed: Binding failed for port 71c05380-b654-4e8a-addc-e438efd4e9ca, please check neutron logs for more information. [ 804.692486] env[62965]: Removing descriptor: 19 [ 804.792856] env[62965]: DEBUG nova.compute.manager [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 804.820652] env[62965]: DEBUG nova.virt.hardware [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 804.820900] env[62965]: DEBUG nova.virt.hardware [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 804.821064] env[62965]: DEBUG nova.virt.hardware [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 804.821247] env[62965]: DEBUG nova.virt.hardware [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 804.821395] env[62965]: DEBUG nova.virt.hardware [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 804.821580] env[62965]: DEBUG nova.virt.hardware [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 804.821803] env[62965]: DEBUG nova.virt.hardware [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 804.821961] env[62965]: DEBUG nova.virt.hardware [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 804.822687] env[62965]: DEBUG nova.virt.hardware [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 804.822940] env[62965]: DEBUG nova.virt.hardware [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 804.823072] env[62965]: DEBUG nova.virt.hardware [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 804.823925] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66c68c30-1a3d-4677-8a57-97683979cc57 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.832663] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c8dc8cd-4bd6-4b4f-a61e-0f952b45caca {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.847969] env[62965]: ERROR nova.compute.manager [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 71c05380-b654-4e8a-addc-e438efd4e9ca, please check neutron logs for more information. [ 804.847969] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Traceback (most recent call last): [ 804.847969] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 804.847969] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] yield resources [ 804.847969] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 804.847969] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] self.driver.spawn(context, instance, image_meta, [ 804.847969] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 804.847969] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 804.847969] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 804.847969] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] vm_ref = self.build_virtual_machine(instance, [ 804.847969] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 804.848304] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] vif_infos = vmwarevif.get_vif_info(self._session, [ 804.848304] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 804.848304] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] for vif in network_info: [ 804.848304] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 804.848304] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] return self._sync_wrapper(fn, *args, **kwargs) [ 804.848304] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 804.848304] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] self.wait() [ 804.848304] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 804.848304] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] self[:] = self._gt.wait() [ 804.848304] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 804.848304] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] return self._exit_event.wait() [ 804.848304] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 804.848304] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] current.throw(*self._exc) [ 804.848657] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 804.848657] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] result = function(*args, **kwargs) [ 804.848657] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 804.848657] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] return func(*args, **kwargs) [ 804.848657] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 804.848657] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] raise e [ 804.848657] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 804.848657] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] nwinfo = self.network_api.allocate_for_instance( [ 804.848657] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 804.848657] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] created_port_ids = self._update_ports_for_instance( [ 804.848657] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 804.848657] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] with excutils.save_and_reraise_exception(): [ 804.848657] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 804.849241] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] self.force_reraise() [ 804.849241] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 804.849241] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] raise self.value [ 804.849241] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 804.849241] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] updated_port = self._update_port( [ 804.849241] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 804.849241] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] _ensure_no_port_binding_failure(port) [ 804.849241] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 804.849241] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] raise exception.PortBindingFailed(port_id=port['id']) [ 804.849241] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] nova.exception.PortBindingFailed: Binding failed for port 71c05380-b654-4e8a-addc-e438efd4e9ca, please check neutron logs for more information. [ 804.849241] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] [ 804.849241] env[62965]: INFO nova.compute.manager [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Terminating instance [ 805.161021] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.409s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.161428] env[62965]: DEBUG nova.compute.manager [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 805.952319] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.578s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.955540] env[62965]: DEBUG nova.network.neutron [-] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.958438] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] Acquiring lock "refresh_cache-ab4da02e-f3ec-473b-968f-14b102780ddb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.958438] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] Acquired lock "refresh_cache-ab4da02e-f3ec-473b-968f-14b102780ddb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.958438] env[62965]: DEBUG nova.network.neutron [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 806.460726] env[62965]: INFO nova.compute.manager [-] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Took 1.81 seconds to deallocate network for instance. [ 806.462566] env[62965]: DEBUG nova.compute.utils [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 806.470053] env[62965]: DEBUG nova.compute.manager [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 806.470053] env[62965]: DEBUG nova.network.neutron [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 806.478204] env[62965]: DEBUG nova.compute.manager [req-68dd4e6d-5aea-4cb7-b948-ab2f822864ae req-73f59d35-8ae8-49b0-be99-fac6145d269f service nova] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Received event network-changed-71c05380-b654-4e8a-addc-e438efd4e9ca {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 806.478204] env[62965]: DEBUG nova.compute.manager [req-68dd4e6d-5aea-4cb7-b948-ab2f822864ae req-73f59d35-8ae8-49b0-be99-fac6145d269f service nova] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Refreshing instance network info cache due to event network-changed-71c05380-b654-4e8a-addc-e438efd4e9ca. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 806.478204] env[62965]: DEBUG oslo_concurrency.lockutils [req-68dd4e6d-5aea-4cb7-b948-ab2f822864ae req-73f59d35-8ae8-49b0-be99-fac6145d269f service nova] Acquiring lock "refresh_cache-ab4da02e-f3ec-473b-968f-14b102780ddb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.478204] env[62965]: DEBUG nova.compute.claims [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 806.478204] env[62965]: DEBUG oslo_concurrency.lockutils [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.496985] env[62965]: DEBUG nova.network.neutron [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 806.557909] env[62965]: DEBUG nova.policy [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '72559d6fef82453b92be892dcc71cec8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4c7fd7540a9b4477a3fa20ce5830009b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 806.630051] env[62965]: DEBUG nova.network.neutron [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.866886] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbae09fb-facf-4389-9cb0-4164047d0c64 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.877094] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-141b696b-a489-4d77-8fff-188a9a4c4a2d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.913023] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-def4490c-21d6-4428-8b9e-b987e6430743 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.921149] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9955521-5e30-41e3-881e-64879d3a233d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.934100] env[62965]: DEBUG nova.compute.provider_tree [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 806.976395] env[62965]: DEBUG nova.compute.manager [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 807.097539] env[62965]: DEBUG nova.network.neutron [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Successfully created port: 72f714ef-8c25-45e9-b59d-636aab193fb5 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 807.133297] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] Releasing lock "refresh_cache-ab4da02e-f3ec-473b-968f-14b102780ddb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.133746] env[62965]: DEBUG nova.compute.manager [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 807.133935] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 807.134262] env[62965]: DEBUG oslo_concurrency.lockutils [req-68dd4e6d-5aea-4cb7-b948-ab2f822864ae req-73f59d35-8ae8-49b0-be99-fac6145d269f service nova] Acquired lock "refresh_cache-ab4da02e-f3ec-473b-968f-14b102780ddb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.134434] env[62965]: DEBUG nova.network.neutron [req-68dd4e6d-5aea-4cb7-b948-ab2f822864ae req-73f59d35-8ae8-49b0-be99-fac6145d269f service nova] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Refreshing network info cache for port 71c05380-b654-4e8a-addc-e438efd4e9ca {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 807.135533] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5dc21015-91fc-48b4-915b-9552f80fa9f2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.150488] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0b9382a-1d15-412a-8c7a-dc1f7bee71ae {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.174550] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ab4da02e-f3ec-473b-968f-14b102780ddb could not be found. [ 807.174780] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 807.174956] env[62965]: INFO nova.compute.manager [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Took 0.04 seconds to destroy the instance on the hypervisor. [ 807.175236] env[62965]: DEBUG oslo.service.loopingcall [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 807.178922] env[62965]: DEBUG nova.compute.manager [-] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 807.178922] env[62965]: DEBUG nova.network.neutron [-] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 807.202831] env[62965]: DEBUG nova.network.neutron [-] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 807.441975] env[62965]: DEBUG nova.scheduler.client.report [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 807.482971] env[62965]: INFO nova.virt.block_device [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Booting with volume a346bc67-ff3f-4650-bc9c-c91d97212609 at /dev/sda [ 807.536077] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-79cda640-3c50-4771-af64-5cbf3871e9f7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.548194] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c7eefaf-752c-4c9c-b66e-a1d838ea061c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.570765] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6f048914-5272-42a4-aec8-e67c6bcf355c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.578533] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee88257f-e174-4d53-848b-32ca9b33f562 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.602576] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc6768db-2062-48d8-b3dd-22bdaaafd2e9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.608396] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26895e9c-9320-4ef3-81c8-241ce4ac0b6b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.621055] env[62965]: DEBUG nova.virt.block_device [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Updating existing volume attachment record: 50e55962-97fa-4542-9d06-f887cab408a8 {{(pid=62965) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 807.669753] env[62965]: DEBUG nova.network.neutron [req-68dd4e6d-5aea-4cb7-b948-ab2f822864ae req-73f59d35-8ae8-49b0-be99-fac6145d269f service nova] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 807.706150] env[62965]: DEBUG nova.network.neutron [-] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.948187] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.996s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.949037] env[62965]: ERROR nova.compute.manager [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d003e564-27fb-48ac-a041-6b3c87abf7ef, please check neutron logs for more information. [ 807.949037] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Traceback (most recent call last): [ 807.949037] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 807.949037] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] self.driver.spawn(context, instance, image_meta, [ 807.949037] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 807.949037] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 807.949037] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 807.949037] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] vm_ref = self.build_virtual_machine(instance, [ 807.949037] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 807.949037] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] vif_infos = vmwarevif.get_vif_info(self._session, [ 807.949037] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 807.949446] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] for vif in network_info: [ 807.949446] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 807.949446] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] return self._sync_wrapper(fn, *args, **kwargs) [ 807.949446] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 807.949446] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] self.wait() [ 807.949446] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 807.949446] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] self[:] = self._gt.wait() [ 807.949446] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 807.949446] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] return self._exit_event.wait() [ 807.949446] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 807.949446] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] current.throw(*self._exc) [ 807.949446] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 807.949446] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] result = function(*args, **kwargs) [ 807.949733] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 807.949733] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] return func(*args, **kwargs) [ 807.949733] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 807.949733] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] raise e [ 807.949733] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 807.949733] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] nwinfo = self.network_api.allocate_for_instance( [ 807.949733] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 807.949733] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] created_port_ids = self._update_ports_for_instance( [ 807.949733] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 807.949733] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] with excutils.save_and_reraise_exception(): [ 807.949733] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 807.949733] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] self.force_reraise() [ 807.949733] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 807.950024] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] raise self.value [ 807.950024] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 807.950024] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] updated_port = self._update_port( [ 807.950024] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 807.950024] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] _ensure_no_port_binding_failure(port) [ 807.950024] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 807.950024] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] raise exception.PortBindingFailed(port_id=port['id']) [ 807.950024] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] nova.exception.PortBindingFailed: Binding failed for port d003e564-27fb-48ac-a041-6b3c87abf7ef, please check neutron logs for more information. [ 807.950024] env[62965]: ERROR nova.compute.manager [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] [ 807.950024] env[62965]: DEBUG nova.compute.utils [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Binding failed for port d003e564-27fb-48ac-a041-6b3c87abf7ef, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 807.954123] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.947s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.957368] env[62965]: INFO nova.compute.claims [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 807.965697] env[62965]: DEBUG nova.compute.manager [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Build of instance dcae9cae-d4f5-4b34-b392-00b5ef9c81bf was re-scheduled: Binding failed for port d003e564-27fb-48ac-a041-6b3c87abf7ef, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 807.965697] env[62965]: DEBUG nova.compute.manager [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 807.965697] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "refresh_cache-dcae9cae-d4f5-4b34-b392-00b5ef9c81bf" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.965697] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquired lock "refresh_cache-dcae9cae-d4f5-4b34-b392-00b5ef9c81bf" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.965903] env[62965]: DEBUG nova.network.neutron [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 807.966898] env[62965]: ERROR nova.compute.manager [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 72f714ef-8c25-45e9-b59d-636aab193fb5, please check neutron logs for more information. [ 807.966898] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 807.966898] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 807.966898] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 807.966898] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 807.966898] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 807.966898] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 807.966898] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 807.966898] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 807.966898] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 807.966898] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 807.966898] env[62965]: ERROR nova.compute.manager raise self.value [ 807.966898] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 807.966898] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 807.966898] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 807.966898] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 807.967350] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 807.967350] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 807.967350] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 72f714ef-8c25-45e9-b59d-636aab193fb5, please check neutron logs for more information. [ 807.967350] env[62965]: ERROR nova.compute.manager [ 807.967350] env[62965]: Traceback (most recent call last): [ 807.967350] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 807.967350] env[62965]: listener.cb(fileno) [ 807.967350] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 807.967350] env[62965]: result = function(*args, **kwargs) [ 807.967350] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 807.967350] env[62965]: return func(*args, **kwargs) [ 807.967350] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 807.967350] env[62965]: raise e [ 807.967350] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 807.967350] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 807.967350] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 807.967350] env[62965]: created_port_ids = self._update_ports_for_instance( [ 807.967350] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 807.967350] env[62965]: with excutils.save_and_reraise_exception(): [ 807.967350] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 807.967350] env[62965]: self.force_reraise() [ 807.967350] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 807.967350] env[62965]: raise self.value [ 807.967849] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 807.967849] env[62965]: updated_port = self._update_port( [ 807.967849] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 807.967849] env[62965]: _ensure_no_port_binding_failure(port) [ 807.967849] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 807.967849] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 807.967849] env[62965]: nova.exception.PortBindingFailed: Binding failed for port 72f714ef-8c25-45e9-b59d-636aab193fb5, please check neutron logs for more information. [ 807.967849] env[62965]: Removing descriptor: 19 [ 807.969965] env[62965]: DEBUG nova.network.neutron [req-68dd4e6d-5aea-4cb7-b948-ab2f822864ae req-73f59d35-8ae8-49b0-be99-fac6145d269f service nova] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.207989] env[62965]: INFO nova.compute.manager [-] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Took 1.03 seconds to deallocate network for instance. [ 808.210536] env[62965]: DEBUG nova.compute.claims [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 808.210719] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.475212] env[62965]: DEBUG oslo_concurrency.lockutils [req-68dd4e6d-5aea-4cb7-b948-ab2f822864ae req-73f59d35-8ae8-49b0-be99-fac6145d269f service nova] Releasing lock "refresh_cache-ab4da02e-f3ec-473b-968f-14b102780ddb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.475212] env[62965]: DEBUG nova.compute.manager [req-68dd4e6d-5aea-4cb7-b948-ab2f822864ae req-73f59d35-8ae8-49b0-be99-fac6145d269f service nova] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Received event network-vif-deleted-71c05380-b654-4e8a-addc-e438efd4e9ca {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 808.488547] env[62965]: DEBUG nova.network.neutron [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 808.492402] env[62965]: DEBUG nova.compute.manager [req-5ec4951d-c3ed-4058-b706-0a5d4a08d110 req-ba97d305-c76b-4b21-850c-59169f11042f service nova] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Received event network-changed-72f714ef-8c25-45e9-b59d-636aab193fb5 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 808.492581] env[62965]: DEBUG nova.compute.manager [req-5ec4951d-c3ed-4058-b706-0a5d4a08d110 req-ba97d305-c76b-4b21-850c-59169f11042f service nova] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Refreshing instance network info cache due to event network-changed-72f714ef-8c25-45e9-b59d-636aab193fb5. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 808.492779] env[62965]: DEBUG oslo_concurrency.lockutils [req-5ec4951d-c3ed-4058-b706-0a5d4a08d110 req-ba97d305-c76b-4b21-850c-59169f11042f service nova] Acquiring lock "refresh_cache-c8402430-4229-4b61-9118-ddeab4d8cd79" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.492916] env[62965]: DEBUG oslo_concurrency.lockutils [req-5ec4951d-c3ed-4058-b706-0a5d4a08d110 req-ba97d305-c76b-4b21-850c-59169f11042f service nova] Acquired lock "refresh_cache-c8402430-4229-4b61-9118-ddeab4d8cd79" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.493080] env[62965]: DEBUG nova.network.neutron [req-5ec4951d-c3ed-4058-b706-0a5d4a08d110 req-ba97d305-c76b-4b21-850c-59169f11042f service nova] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Refreshing network info cache for port 72f714ef-8c25-45e9-b59d-636aab193fb5 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 808.575353] env[62965]: DEBUG nova.network.neutron [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.701869] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Acquiring lock "924c5cdc-0450-4ced-b920-50e2d5060fd2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.702222] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Lock "924c5cdc-0450-4ced-b920-50e2d5060fd2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.014304] env[62965]: DEBUG nova.network.neutron [req-5ec4951d-c3ed-4058-b706-0a5d4a08d110 req-ba97d305-c76b-4b21-850c-59169f11042f service nova] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 809.078173] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Releasing lock "refresh_cache-dcae9cae-d4f5-4b34-b392-00b5ef9c81bf" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.078406] env[62965]: DEBUG nova.compute.manager [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 809.078578] env[62965]: DEBUG nova.compute.manager [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 809.078755] env[62965]: DEBUG nova.network.neutron [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 809.093700] env[62965]: DEBUG nova.network.neutron [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 809.111728] env[62965]: DEBUG nova.network.neutron [req-5ec4951d-c3ed-4058-b706-0a5d4a08d110 req-ba97d305-c76b-4b21-850c-59169f11042f service nova] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.202824] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-990176c8-80f1-4125-ad0a-d8911215e793 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.211645] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a85ebb6-276d-42bc-a49d-a7a06ce0297a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.242762] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85206f7f-e5d7-4e77-962e-dbf6aaa37f45 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.250733] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9d9c53e-dce0-4de4-bd36-1b09eb614013 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.265241] env[62965]: DEBUG nova.compute.provider_tree [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 809.598946] env[62965]: DEBUG nova.network.neutron [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.614224] env[62965]: DEBUG oslo_concurrency.lockutils [req-5ec4951d-c3ed-4058-b706-0a5d4a08d110 req-ba97d305-c76b-4b21-850c-59169f11042f service nova] Releasing lock "refresh_cache-c8402430-4229-4b61-9118-ddeab4d8cd79" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.614479] env[62965]: DEBUG nova.compute.manager [req-5ec4951d-c3ed-4058-b706-0a5d4a08d110 req-ba97d305-c76b-4b21-850c-59169f11042f service nova] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Received event network-vif-deleted-72f714ef-8c25-45e9-b59d-636aab193fb5 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 809.744143] env[62965]: DEBUG nova.compute.manager [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 809.744684] env[62965]: DEBUG nova.virt.hardware [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 809.744895] env[62965]: DEBUG nova.virt.hardware [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 809.745062] env[62965]: DEBUG nova.virt.hardware [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 809.745252] env[62965]: DEBUG nova.virt.hardware [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 809.745398] env[62965]: DEBUG nova.virt.hardware [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 809.745543] env[62965]: DEBUG nova.virt.hardware [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 809.745745] env[62965]: DEBUG nova.virt.hardware [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 809.745901] env[62965]: DEBUG nova.virt.hardware [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 809.746082] env[62965]: DEBUG nova.virt.hardware [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 809.746248] env[62965]: DEBUG nova.virt.hardware [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 809.746420] env[62965]: DEBUG nova.virt.hardware [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 809.747356] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22e08130-8a87-465a-8365-15579dfae22c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.755862] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-594e05d5-a73d-421e-8184-4e7fc94c1484 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.769195] env[62965]: DEBUG nova.scheduler.client.report [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 809.772698] env[62965]: ERROR nova.compute.manager [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 72f714ef-8c25-45e9-b59d-636aab193fb5, please check neutron logs for more information. [ 809.772698] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Traceback (most recent call last): [ 809.772698] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 809.772698] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] yield resources [ 809.772698] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 809.772698] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] self.driver.spawn(context, instance, image_meta, [ 809.772698] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 809.772698] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] self._vmops.spawn(context, instance, image_meta, injected_files, [ 809.772698] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 809.772698] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] vm_ref = self.build_virtual_machine(instance, [ 809.772698] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 809.773067] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] vif_infos = vmwarevif.get_vif_info(self._session, [ 809.773067] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 809.773067] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] for vif in network_info: [ 809.773067] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 809.773067] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] return self._sync_wrapper(fn, *args, **kwargs) [ 809.773067] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 809.773067] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] self.wait() [ 809.773067] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 809.773067] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] self[:] = self._gt.wait() [ 809.773067] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 809.773067] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] return self._exit_event.wait() [ 809.773067] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 809.773067] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] current.throw(*self._exc) [ 809.773539] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 809.773539] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] result = function(*args, **kwargs) [ 809.773539] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 809.773539] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] return func(*args, **kwargs) [ 809.773539] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 809.773539] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] raise e [ 809.773539] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 809.773539] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] nwinfo = self.network_api.allocate_for_instance( [ 809.773539] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 809.773539] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] created_port_ids = self._update_ports_for_instance( [ 809.773539] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 809.773539] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] with excutils.save_and_reraise_exception(): [ 809.773539] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 809.773891] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] self.force_reraise() [ 809.773891] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 809.773891] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] raise self.value [ 809.773891] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 809.773891] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] updated_port = self._update_port( [ 809.773891] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 809.773891] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] _ensure_no_port_binding_failure(port) [ 809.773891] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 809.773891] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] raise exception.PortBindingFailed(port_id=port['id']) [ 809.773891] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] nova.exception.PortBindingFailed: Binding failed for port 72f714ef-8c25-45e9-b59d-636aab193fb5, please check neutron logs for more information. [ 809.773891] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] [ 809.773891] env[62965]: INFO nova.compute.manager [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Terminating instance [ 810.102441] env[62965]: INFO nova.compute.manager [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: dcae9cae-d4f5-4b34-b392-00b5ef9c81bf] Took 1.02 seconds to deallocate network for instance. [ 810.276220] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.324s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.276742] env[62965]: DEBUG nova.compute.manager [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 810.279436] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] Acquiring lock "refresh_cache-c8402430-4229-4b61-9118-ddeab4d8cd79" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.279594] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] Acquired lock "refresh_cache-c8402430-4229-4b61-9118-ddeab4d8cd79" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.279757] env[62965]: DEBUG nova.network.neutron [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 810.281233] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.425s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.282729] env[62965]: INFO nova.compute.claims [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 810.782946] env[62965]: DEBUG nova.compute.utils [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 810.784308] env[62965]: DEBUG nova.compute.manager [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 810.784482] env[62965]: DEBUG nova.network.neutron [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 810.807027] env[62965]: DEBUG nova.network.neutron [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 810.841173] env[62965]: DEBUG nova.policy [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4aa5b297d4e04d2f82a8d02d7559fa3d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2f18fd3a91f242f3b232b9fd6b985a43', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 810.877165] env[62965]: DEBUG nova.network.neutron [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.108413] env[62965]: DEBUG nova.network.neutron [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Successfully created port: 9c95488d-c11e-4946-b03a-1d15547e81a1 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 811.134605] env[62965]: INFO nova.scheduler.client.report [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Deleted allocations for instance dcae9cae-d4f5-4b34-b392-00b5ef9c81bf [ 811.288041] env[62965]: DEBUG nova.compute.manager [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 811.380342] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] Releasing lock "refresh_cache-c8402430-4229-4b61-9118-ddeab4d8cd79" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.380894] env[62965]: DEBUG nova.compute.manager [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 811.381247] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-08d3b84f-ffa1-4e07-be4d-f050a4f2fe7f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.391012] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0826b925-3b40-44da-8165-8234d4497fa9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.416019] env[62965]: WARNING nova.virt.vmwareapi.driver [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance c8402430-4229-4b61-9118-ddeab4d8cd79 could not be found. [ 811.416019] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 811.416019] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1f1e9c8c-33e6-47e9-b61b-754161051f20 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.423560] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbb2ea11-ec5b-48e7-9a52-37c6df4989d0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.448466] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c8402430-4229-4b61-9118-ddeab4d8cd79 could not be found. [ 811.449362] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 811.449763] env[62965]: INFO nova.compute.manager [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Took 0.07 seconds to destroy the instance on the hypervisor. [ 811.450227] env[62965]: DEBUG oslo.service.loopingcall [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 811.450635] env[62965]: DEBUG nova.compute.manager [-] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 811.452457] env[62965]: DEBUG nova.network.neutron [-] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 811.473256] env[62965]: DEBUG nova.network.neutron [-] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 811.628359] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c403f7ef-5d27-4fcf-bd81-29a883239982 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.637246] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f8b9671-50e3-4a7a-9945-79b3f66f9a3b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.643666] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c41a5dd-9a84-4234-805c-43aed5756056 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "dcae9cae-d4f5-4b34-b392-00b5ef9c81bf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.198s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.679556] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b850b21-7755-4dca-bf11-db924d4d4845 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.690014] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18c5f7f2-4a33-4817-a510-05fe5fccbe70 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.704249] env[62965]: DEBUG nova.compute.provider_tree [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 811.974494] env[62965]: DEBUG nova.network.neutron [-] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.014671] env[62965]: DEBUG nova.compute.manager [req-55e1c3f8-19f7-4a1a-bfe3-764f16045851 req-00422b7e-fd62-473b-9ee6-a6defd88092a service nova] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Received event network-changed-9c95488d-c11e-4946-b03a-1d15547e81a1 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 812.017286] env[62965]: DEBUG nova.compute.manager [req-55e1c3f8-19f7-4a1a-bfe3-764f16045851 req-00422b7e-fd62-473b-9ee6-a6defd88092a service nova] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Refreshing instance network info cache due to event network-changed-9c95488d-c11e-4946-b03a-1d15547e81a1. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 812.017952] env[62965]: DEBUG oslo_concurrency.lockutils [req-55e1c3f8-19f7-4a1a-bfe3-764f16045851 req-00422b7e-fd62-473b-9ee6-a6defd88092a service nova] Acquiring lock "refresh_cache-a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.018347] env[62965]: DEBUG oslo_concurrency.lockutils [req-55e1c3f8-19f7-4a1a-bfe3-764f16045851 req-00422b7e-fd62-473b-9ee6-a6defd88092a service nova] Acquired lock "refresh_cache-a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.018579] env[62965]: DEBUG nova.network.neutron [req-55e1c3f8-19f7-4a1a-bfe3-764f16045851 req-00422b7e-fd62-473b-9ee6-a6defd88092a service nova] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Refreshing network info cache for port 9c95488d-c11e-4946-b03a-1d15547e81a1 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 812.185118] env[62965]: DEBUG nova.compute.manager [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 812.207401] env[62965]: DEBUG nova.scheduler.client.report [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 812.251941] env[62965]: ERROR nova.compute.manager [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9c95488d-c11e-4946-b03a-1d15547e81a1, please check neutron logs for more information. [ 812.251941] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 812.251941] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 812.251941] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 812.251941] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 812.251941] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 812.251941] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 812.251941] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 812.251941] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 812.251941] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 812.251941] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 812.251941] env[62965]: ERROR nova.compute.manager raise self.value [ 812.251941] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 812.251941] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 812.251941] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 812.251941] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 812.252367] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 812.252367] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 812.252367] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9c95488d-c11e-4946-b03a-1d15547e81a1, please check neutron logs for more information. [ 812.252367] env[62965]: ERROR nova.compute.manager [ 812.252367] env[62965]: Traceback (most recent call last): [ 812.252367] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 812.252367] env[62965]: listener.cb(fileno) [ 812.252367] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 812.252367] env[62965]: result = function(*args, **kwargs) [ 812.252367] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 812.252367] env[62965]: return func(*args, **kwargs) [ 812.252367] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 812.252367] env[62965]: raise e [ 812.252367] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 812.252367] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 812.252367] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 812.252367] env[62965]: created_port_ids = self._update_ports_for_instance( [ 812.252367] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 812.252367] env[62965]: with excutils.save_and_reraise_exception(): [ 812.252367] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 812.252367] env[62965]: self.force_reraise() [ 812.252367] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 812.252367] env[62965]: raise self.value [ 812.252367] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 812.252367] env[62965]: updated_port = self._update_port( [ 812.252367] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 812.252367] env[62965]: _ensure_no_port_binding_failure(port) [ 812.252367] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 812.252367] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 812.252985] env[62965]: nova.exception.PortBindingFailed: Binding failed for port 9c95488d-c11e-4946-b03a-1d15547e81a1, please check neutron logs for more information. [ 812.252985] env[62965]: Removing descriptor: 17 [ 812.303679] env[62965]: DEBUG nova.compute.manager [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 812.332805] env[62965]: DEBUG nova.virt.hardware [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 812.333328] env[62965]: DEBUG nova.virt.hardware [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 812.333551] env[62965]: DEBUG nova.virt.hardware [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 812.333947] env[62965]: DEBUG nova.virt.hardware [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 812.334364] env[62965]: DEBUG nova.virt.hardware [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 812.334570] env[62965]: DEBUG nova.virt.hardware [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 812.334824] env[62965]: DEBUG nova.virt.hardware [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 812.335120] env[62965]: DEBUG nova.virt.hardware [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 812.335357] env[62965]: DEBUG nova.virt.hardware [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 812.335568] env[62965]: DEBUG nova.virt.hardware [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 812.335779] env[62965]: DEBUG nova.virt.hardware [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 812.337550] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0d1b9b3-2974-44b7-87d0-438e56e8da95 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.346320] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b46ce85-40c3-47c6-9a8b-37b65f7fc7ab {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.360927] env[62965]: ERROR nova.compute.manager [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9c95488d-c11e-4946-b03a-1d15547e81a1, please check neutron logs for more information. [ 812.360927] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Traceback (most recent call last): [ 812.360927] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 812.360927] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] yield resources [ 812.360927] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 812.360927] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] self.driver.spawn(context, instance, image_meta, [ 812.360927] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 812.360927] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 812.360927] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 812.360927] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] vm_ref = self.build_virtual_machine(instance, [ 812.360927] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 812.361234] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] vif_infos = vmwarevif.get_vif_info(self._session, [ 812.361234] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 812.361234] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] for vif in network_info: [ 812.361234] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 812.361234] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] return self._sync_wrapper(fn, *args, **kwargs) [ 812.361234] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 812.361234] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] self.wait() [ 812.361234] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 812.361234] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] self[:] = self._gt.wait() [ 812.361234] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 812.361234] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] return self._exit_event.wait() [ 812.361234] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 812.361234] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] current.throw(*self._exc) [ 812.361517] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 812.361517] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] result = function(*args, **kwargs) [ 812.361517] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 812.361517] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] return func(*args, **kwargs) [ 812.361517] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 812.361517] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] raise e [ 812.361517] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 812.361517] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] nwinfo = self.network_api.allocate_for_instance( [ 812.361517] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 812.361517] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] created_port_ids = self._update_ports_for_instance( [ 812.361517] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 812.361517] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] with excutils.save_and_reraise_exception(): [ 812.361517] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 812.361796] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] self.force_reraise() [ 812.361796] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 812.361796] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] raise self.value [ 812.361796] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 812.361796] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] updated_port = self._update_port( [ 812.361796] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 812.361796] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] _ensure_no_port_binding_failure(port) [ 812.361796] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 812.361796] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] raise exception.PortBindingFailed(port_id=port['id']) [ 812.361796] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] nova.exception.PortBindingFailed: Binding failed for port 9c95488d-c11e-4946-b03a-1d15547e81a1, please check neutron logs for more information. [ 812.361796] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] [ 812.361796] env[62965]: INFO nova.compute.manager [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Terminating instance [ 812.476923] env[62965]: INFO nova.compute.manager [-] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Took 1.03 seconds to deallocate network for instance. [ 812.541079] env[62965]: DEBUG nova.network.neutron [req-55e1c3f8-19f7-4a1a-bfe3-764f16045851 req-00422b7e-fd62-473b-9ee6-a6defd88092a service nova] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 812.612566] env[62965]: DEBUG nova.network.neutron [req-55e1c3f8-19f7-4a1a-bfe3-764f16045851 req-00422b7e-fd62-473b-9ee6-a6defd88092a service nova] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.704765] env[62965]: DEBUG oslo_concurrency.lockutils [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.714636] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.433s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.715102] env[62965]: DEBUG nova.compute.manager [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 812.717750] env[62965]: DEBUG oslo_concurrency.lockutils [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.035s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.719489] env[62965]: INFO nova.compute.claims [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 812.865873] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] Acquiring lock "refresh_cache-a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.019455] env[62965]: DEBUG oslo_concurrency.lockutils [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "8c478341-9b50-45de-9ba3-321e802ed1f0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.019654] env[62965]: DEBUG oslo_concurrency.lockutils [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "8c478341-9b50-45de-9ba3-321e802ed1f0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.033014] env[62965]: INFO nova.compute.manager [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Took 0.56 seconds to detach 1 volumes for instance. [ 813.034984] env[62965]: DEBUG nova.compute.claims [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 813.035180] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.114564] env[62965]: DEBUG oslo_concurrency.lockutils [req-55e1c3f8-19f7-4a1a-bfe3-764f16045851 req-00422b7e-fd62-473b-9ee6-a6defd88092a service nova] Releasing lock "refresh_cache-a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.114943] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] Acquired lock "refresh_cache-a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.115152] env[62965]: DEBUG nova.network.neutron [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 813.223981] env[62965]: DEBUG nova.compute.utils [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 813.227246] env[62965]: DEBUG nova.compute.manager [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 813.227415] env[62965]: DEBUG nova.network.neutron [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 813.265215] env[62965]: DEBUG nova.policy [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a5e5c6c3198946658126b0024ba82583', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9a539017d4b84f39a71b6f4098395fbe', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 813.624198] env[62965]: DEBUG nova.network.neutron [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Successfully created port: feea997a-7e87-460c-93a4-134225451c55 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 813.634462] env[62965]: DEBUG nova.network.neutron [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 813.706928] env[62965]: DEBUG nova.network.neutron [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.727904] env[62965]: DEBUG nova.compute.manager [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 813.990246] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4977c6f-6787-4c60-a5e0-cbe4065c73bc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.999599] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d08ea08f-fee1-4220-b452-80613c2b3233 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.027214] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c3a0a0e-36fc-43c3-9a69-f88b7e2ef5d0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.034898] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16da84f9-543d-497b-bfe4-24cf29fc7dd0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.041055] env[62965]: DEBUG nova.compute.manager [req-9db557ea-3118-474e-907d-26a56c908075 req-276c6a73-d68e-4367-8525-29b523847f95 service nova] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Received event network-vif-deleted-9c95488d-c11e-4946-b03a-1d15547e81a1 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 814.050294] env[62965]: DEBUG nova.compute.provider_tree [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 814.211585] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] Releasing lock "refresh_cache-a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.215231] env[62965]: DEBUG nova.compute.manager [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 814.215447] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 814.215753] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9c51082b-16df-4ec5-9d45-adbbe1d93de6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.224625] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfcad9a3-8688-466b-ac38-c009818d8882 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.255362] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7 could not be found. [ 814.255577] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 814.255755] env[62965]: INFO nova.compute.manager [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Took 0.04 seconds to destroy the instance on the hypervisor. [ 814.256019] env[62965]: DEBUG oslo.service.loopingcall [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 814.256261] env[62965]: DEBUG nova.compute.manager [-] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 814.256357] env[62965]: DEBUG nova.network.neutron [-] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 814.271251] env[62965]: DEBUG nova.network.neutron [-] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 814.554954] env[62965]: DEBUG nova.scheduler.client.report [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 814.628470] env[62965]: ERROR nova.compute.manager [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port feea997a-7e87-460c-93a4-134225451c55, please check neutron logs for more information. [ 814.628470] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 814.628470] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 814.628470] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 814.628470] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 814.628470] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 814.628470] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 814.628470] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 814.628470] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 814.628470] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 814.628470] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 814.628470] env[62965]: ERROR nova.compute.manager raise self.value [ 814.628470] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 814.628470] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 814.628470] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 814.628470] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 814.629114] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 814.629114] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 814.629114] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port feea997a-7e87-460c-93a4-134225451c55, please check neutron logs for more information. [ 814.629114] env[62965]: ERROR nova.compute.manager [ 814.629114] env[62965]: Traceback (most recent call last): [ 814.629114] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 814.629114] env[62965]: listener.cb(fileno) [ 814.629114] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 814.629114] env[62965]: result = function(*args, **kwargs) [ 814.629114] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 814.629114] env[62965]: return func(*args, **kwargs) [ 814.629114] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 814.629114] env[62965]: raise e [ 814.629114] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 814.629114] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 814.629114] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 814.629114] env[62965]: created_port_ids = self._update_ports_for_instance( [ 814.629114] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 814.629114] env[62965]: with excutils.save_and_reraise_exception(): [ 814.629114] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 814.629114] env[62965]: self.force_reraise() [ 814.629114] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 814.629114] env[62965]: raise self.value [ 814.629114] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 814.629114] env[62965]: updated_port = self._update_port( [ 814.629114] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 814.629114] env[62965]: _ensure_no_port_binding_failure(port) [ 814.629114] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 814.629114] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 814.629854] env[62965]: nova.exception.PortBindingFailed: Binding failed for port feea997a-7e87-460c-93a4-134225451c55, please check neutron logs for more information. [ 814.629854] env[62965]: Removing descriptor: 17 [ 814.746136] env[62965]: DEBUG nova.compute.manager [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 814.769725] env[62965]: DEBUG nova.virt.hardware [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 814.769996] env[62965]: DEBUG nova.virt.hardware [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 814.770187] env[62965]: DEBUG nova.virt.hardware [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 814.770395] env[62965]: DEBUG nova.virt.hardware [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 814.770538] env[62965]: DEBUG nova.virt.hardware [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 814.770681] env[62965]: DEBUG nova.virt.hardware [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 814.770881] env[62965]: DEBUG nova.virt.hardware [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 814.771044] env[62965]: DEBUG nova.virt.hardware [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 814.771213] env[62965]: DEBUG nova.virt.hardware [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 814.771372] env[62965]: DEBUG nova.virt.hardware [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 814.771540] env[62965]: DEBUG nova.virt.hardware [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 814.772382] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-181f9200-4ca1-4707-9eec-702574f1e14b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.775008] env[62965]: DEBUG nova.network.neutron [-] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 814.781121] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5f3ac6b-6285-4833-b630-673b288028c6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.795904] env[62965]: ERROR nova.compute.manager [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port feea997a-7e87-460c-93a4-134225451c55, please check neutron logs for more information. [ 814.795904] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Traceback (most recent call last): [ 814.795904] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 814.795904] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] yield resources [ 814.795904] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 814.795904] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] self.driver.spawn(context, instance, image_meta, [ 814.795904] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 814.795904] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 814.795904] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 814.795904] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] vm_ref = self.build_virtual_machine(instance, [ 814.795904] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 814.796236] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] vif_infos = vmwarevif.get_vif_info(self._session, [ 814.796236] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 814.796236] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] for vif in network_info: [ 814.796236] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 814.796236] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] return self._sync_wrapper(fn, *args, **kwargs) [ 814.796236] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 814.796236] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] self.wait() [ 814.796236] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 814.796236] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] self[:] = self._gt.wait() [ 814.796236] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 814.796236] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] return self._exit_event.wait() [ 814.796236] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 814.796236] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] current.throw(*self._exc) [ 814.796530] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 814.796530] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] result = function(*args, **kwargs) [ 814.796530] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 814.796530] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] return func(*args, **kwargs) [ 814.796530] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 814.796530] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] raise e [ 814.796530] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 814.796530] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] nwinfo = self.network_api.allocate_for_instance( [ 814.796530] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 814.796530] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] created_port_ids = self._update_ports_for_instance( [ 814.796530] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 814.796530] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] with excutils.save_and_reraise_exception(): [ 814.796530] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 814.796814] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] self.force_reraise() [ 814.796814] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 814.796814] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] raise self.value [ 814.796814] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 814.796814] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] updated_port = self._update_port( [ 814.796814] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 814.796814] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] _ensure_no_port_binding_failure(port) [ 814.796814] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 814.796814] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] raise exception.PortBindingFailed(port_id=port['id']) [ 814.796814] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] nova.exception.PortBindingFailed: Binding failed for port feea997a-7e87-460c-93a4-134225451c55, please check neutron logs for more information. [ 814.796814] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] [ 814.796814] env[62965]: INFO nova.compute.manager [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Terminating instance [ 815.059991] env[62965]: DEBUG oslo_concurrency.lockutils [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.342s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.060527] env[62965]: DEBUG nova.compute.manager [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 815.063432] env[62965]: DEBUG oslo_concurrency.lockutils [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 28.180s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.063643] env[62965]: DEBUG nova.objects.instance [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62965) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 815.277767] env[62965]: INFO nova.compute.manager [-] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Took 1.02 seconds to deallocate network for instance. [ 815.280272] env[62965]: DEBUG nova.compute.claims [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 815.280476] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.299134] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquiring lock "refresh_cache-21bf78fb-4da7-4484-9678-b22a6d944c0b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.299384] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquired lock "refresh_cache-21bf78fb-4da7-4484-9678-b22a6d944c0b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.299561] env[62965]: DEBUG nova.network.neutron [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 815.570634] env[62965]: DEBUG nova.compute.utils [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 815.571979] env[62965]: DEBUG nova.compute.manager [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 815.572167] env[62965]: DEBUG nova.network.neutron [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 815.637169] env[62965]: DEBUG nova.policy [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '464c651fad19417ba295bfc289b9b81a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c7dd28b699a44d0693c193dcf0b6cd3b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 815.818376] env[62965]: DEBUG nova.network.neutron [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 815.892340] env[62965]: DEBUG nova.network.neutron [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.077034] env[62965]: DEBUG oslo_concurrency.lockutils [None req-64e16f50-b3a0-41f7-963c-2ccc4c33fbcd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.077034] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.463s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.077476] env[62965]: INFO nova.compute.claims [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 816.081161] env[62965]: DEBUG nova.compute.manager [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 816.104134] env[62965]: DEBUG nova.network.neutron [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Successfully created port: 6d23ec55-580a-4388-b3ad-9bfa3d9b47bb {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 816.123289] env[62965]: DEBUG nova.compute.manager [req-cb011501-fd9d-4390-959d-b4889f83efad req-a7b70304-e973-4652-b5b4-63a94e485867 service nova] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Received event network-changed-feea997a-7e87-460c-93a4-134225451c55 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 816.123289] env[62965]: DEBUG nova.compute.manager [req-cb011501-fd9d-4390-959d-b4889f83efad req-a7b70304-e973-4652-b5b4-63a94e485867 service nova] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Refreshing instance network info cache due to event network-changed-feea997a-7e87-460c-93a4-134225451c55. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 816.123289] env[62965]: DEBUG oslo_concurrency.lockutils [req-cb011501-fd9d-4390-959d-b4889f83efad req-a7b70304-e973-4652-b5b4-63a94e485867 service nova] Acquiring lock "refresh_cache-21bf78fb-4da7-4484-9678-b22a6d944c0b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.394389] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Releasing lock "refresh_cache-21bf78fb-4da7-4484-9678-b22a6d944c0b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.394834] env[62965]: DEBUG nova.compute.manager [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 816.395037] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 816.395348] env[62965]: DEBUG oslo_concurrency.lockutils [req-cb011501-fd9d-4390-959d-b4889f83efad req-a7b70304-e973-4652-b5b4-63a94e485867 service nova] Acquired lock "refresh_cache-21bf78fb-4da7-4484-9678-b22a6d944c0b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.395522] env[62965]: DEBUG nova.network.neutron [req-cb011501-fd9d-4390-959d-b4889f83efad req-a7b70304-e973-4652-b5b4-63a94e485867 service nova] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Refreshing network info cache for port feea997a-7e87-460c-93a4-134225451c55 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 816.396582] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5cdc0b64-04a4-43c8-9f1f-097afe11155e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.406457] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0a73a86-cd1c-4be1-88a4-59603d7f69fe {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.430854] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 21bf78fb-4da7-4484-9678-b22a6d944c0b could not be found. [ 816.431078] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 816.431260] env[62965]: INFO nova.compute.manager [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 816.431492] env[62965]: DEBUG oslo.service.loopingcall [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 816.431705] env[62965]: DEBUG nova.compute.manager [-] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 816.431806] env[62965]: DEBUG nova.network.neutron [-] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 816.445981] env[62965]: DEBUG nova.network.neutron [-] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 816.919481] env[62965]: DEBUG nova.network.neutron [req-cb011501-fd9d-4390-959d-b4889f83efad req-a7b70304-e973-4652-b5b4-63a94e485867 service nova] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 816.948775] env[62965]: DEBUG nova.network.neutron [-] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.994339] env[62965]: DEBUG nova.network.neutron [req-cb011501-fd9d-4390-959d-b4889f83efad req-a7b70304-e973-4652-b5b4-63a94e485867 service nova] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.069640] env[62965]: ERROR nova.compute.manager [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6d23ec55-580a-4388-b3ad-9bfa3d9b47bb, please check neutron logs for more information. [ 817.069640] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 817.069640] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 817.069640] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 817.069640] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 817.069640] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 817.069640] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 817.069640] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 817.069640] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 817.069640] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 817.069640] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 817.069640] env[62965]: ERROR nova.compute.manager raise self.value [ 817.069640] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 817.069640] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 817.069640] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 817.069640] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 817.070097] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 817.070097] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 817.070097] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6d23ec55-580a-4388-b3ad-9bfa3d9b47bb, please check neutron logs for more information. [ 817.070097] env[62965]: ERROR nova.compute.manager [ 817.070097] env[62965]: Traceback (most recent call last): [ 817.070097] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 817.070097] env[62965]: listener.cb(fileno) [ 817.070097] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 817.070097] env[62965]: result = function(*args, **kwargs) [ 817.070097] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 817.070097] env[62965]: return func(*args, **kwargs) [ 817.070097] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 817.070097] env[62965]: raise e [ 817.070097] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 817.070097] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 817.070097] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 817.070097] env[62965]: created_port_ids = self._update_ports_for_instance( [ 817.070097] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 817.070097] env[62965]: with excutils.save_and_reraise_exception(): [ 817.070097] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 817.070097] env[62965]: self.force_reraise() [ 817.070097] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 817.070097] env[62965]: raise self.value [ 817.070097] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 817.070097] env[62965]: updated_port = self._update_port( [ 817.070097] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 817.070097] env[62965]: _ensure_no_port_binding_failure(port) [ 817.070097] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 817.070097] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 817.070858] env[62965]: nova.exception.PortBindingFailed: Binding failed for port 6d23ec55-580a-4388-b3ad-9bfa3d9b47bb, please check neutron logs for more information. [ 817.070858] env[62965]: Removing descriptor: 17 [ 817.093152] env[62965]: DEBUG nova.compute.manager [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 817.120643] env[62965]: DEBUG nova.virt.hardware [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 817.120891] env[62965]: DEBUG nova.virt.hardware [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 817.121070] env[62965]: DEBUG nova.virt.hardware [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 817.121281] env[62965]: DEBUG nova.virt.hardware [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 817.121485] env[62965]: DEBUG nova.virt.hardware [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 817.121645] env[62965]: DEBUG nova.virt.hardware [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 817.121850] env[62965]: DEBUG nova.virt.hardware [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 817.122015] env[62965]: DEBUG nova.virt.hardware [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 817.122197] env[62965]: DEBUG nova.virt.hardware [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 817.122359] env[62965]: DEBUG nova.virt.hardware [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 817.122527] env[62965]: DEBUG nova.virt.hardware [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 817.123425] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5d96ada-d456-45f4-9bcb-10b250c6e1af {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.133393] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57c4a464-399d-4e1d-8ca7-c87931b90e13 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.149767] env[62965]: ERROR nova.compute.manager [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6d23ec55-580a-4388-b3ad-9bfa3d9b47bb, please check neutron logs for more information. [ 817.149767] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Traceback (most recent call last): [ 817.149767] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 817.149767] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] yield resources [ 817.149767] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 817.149767] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] self.driver.spawn(context, instance, image_meta, [ 817.149767] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 817.149767] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 817.149767] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 817.149767] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] vm_ref = self.build_virtual_machine(instance, [ 817.149767] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 817.150081] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] vif_infos = vmwarevif.get_vif_info(self._session, [ 817.150081] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 817.150081] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] for vif in network_info: [ 817.150081] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 817.150081] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] return self._sync_wrapper(fn, *args, **kwargs) [ 817.150081] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 817.150081] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] self.wait() [ 817.150081] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 817.150081] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] self[:] = self._gt.wait() [ 817.150081] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 817.150081] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] return self._exit_event.wait() [ 817.150081] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 817.150081] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] current.throw(*self._exc) [ 817.150364] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 817.150364] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] result = function(*args, **kwargs) [ 817.150364] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 817.150364] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] return func(*args, **kwargs) [ 817.150364] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 817.150364] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] raise e [ 817.150364] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 817.150364] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] nwinfo = self.network_api.allocate_for_instance( [ 817.150364] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 817.150364] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] created_port_ids = self._update_ports_for_instance( [ 817.150364] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 817.150364] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] with excutils.save_and_reraise_exception(): [ 817.150364] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 817.150663] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] self.force_reraise() [ 817.150663] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 817.150663] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] raise self.value [ 817.150663] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 817.150663] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] updated_port = self._update_port( [ 817.150663] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 817.150663] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] _ensure_no_port_binding_failure(port) [ 817.150663] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 817.150663] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] raise exception.PortBindingFailed(port_id=port['id']) [ 817.150663] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] nova.exception.PortBindingFailed: Binding failed for port 6d23ec55-580a-4388-b3ad-9bfa3d9b47bb, please check neutron logs for more information. [ 817.150663] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] [ 817.150663] env[62965]: INFO nova.compute.manager [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Terminating instance [ 817.325740] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-646f3363-2314-4921-9dea-17e9188af932 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.332987] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63041d89-59b9-4820-bb3b-6bad36057c77 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.365448] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ac5f792-b39c-4034-9743-6bfebbf11598 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.372553] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1bfae38-47e5-4c54-93c5-9e3f4bb6ea4e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.385248] env[62965]: DEBUG nova.compute.provider_tree [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 817.451476] env[62965]: INFO nova.compute.manager [-] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Took 1.02 seconds to deallocate network for instance. [ 817.453757] env[62965]: DEBUG nova.compute.claims [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 817.454043] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.496966] env[62965]: DEBUG oslo_concurrency.lockutils [req-cb011501-fd9d-4390-959d-b4889f83efad req-a7b70304-e973-4652-b5b4-63a94e485867 service nova] Releasing lock "refresh_cache-21bf78fb-4da7-4484-9678-b22a6d944c0b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.497246] env[62965]: DEBUG nova.compute.manager [req-cb011501-fd9d-4390-959d-b4889f83efad req-a7b70304-e973-4652-b5b4-63a94e485867 service nova] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Received event network-vif-deleted-feea997a-7e87-460c-93a4-134225451c55 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 817.657314] env[62965]: DEBUG oslo_concurrency.lockutils [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] Acquiring lock "refresh_cache-13d20f82-fa4f-41ab-b08f-5b64da7dddc0" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.657529] env[62965]: DEBUG oslo_concurrency.lockutils [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] Acquired lock "refresh_cache-13d20f82-fa4f-41ab-b08f-5b64da7dddc0" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.657707] env[62965]: DEBUG nova.network.neutron [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 817.888668] env[62965]: DEBUG nova.scheduler.client.report [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 818.163520] env[62965]: DEBUG nova.compute.manager [req-93a928bd-c56e-4b0e-a644-02c4938fd7f2 req-f838dd22-f0a8-484b-a6b9-2bd8553a6ed3 service nova] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Received event network-changed-6d23ec55-580a-4388-b3ad-9bfa3d9b47bb {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 818.163758] env[62965]: DEBUG nova.compute.manager [req-93a928bd-c56e-4b0e-a644-02c4938fd7f2 req-f838dd22-f0a8-484b-a6b9-2bd8553a6ed3 service nova] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Refreshing instance network info cache due to event network-changed-6d23ec55-580a-4388-b3ad-9bfa3d9b47bb. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 818.163949] env[62965]: DEBUG oslo_concurrency.lockutils [req-93a928bd-c56e-4b0e-a644-02c4938fd7f2 req-f838dd22-f0a8-484b-a6b9-2bd8553a6ed3 service nova] Acquiring lock "refresh_cache-13d20f82-fa4f-41ab-b08f-5b64da7dddc0" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.177268] env[62965]: DEBUG nova.network.neutron [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 818.300986] env[62965]: DEBUG nova.network.neutron [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.393370] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.319s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.393887] env[62965]: DEBUG nova.compute.manager [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 818.396364] env[62965]: DEBUG oslo_concurrency.lockutils [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.699s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.396584] env[62965]: DEBUG nova.objects.instance [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Lazy-loading 'resources' on Instance uuid e87adfae-3ea4-4cc4-9978-d8c811c80941 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 818.803760] env[62965]: DEBUG oslo_concurrency.lockutils [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] Releasing lock "refresh_cache-13d20f82-fa4f-41ab-b08f-5b64da7dddc0" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.804174] env[62965]: DEBUG nova.compute.manager [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 818.804370] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 818.804708] env[62965]: DEBUG oslo_concurrency.lockutils [req-93a928bd-c56e-4b0e-a644-02c4938fd7f2 req-f838dd22-f0a8-484b-a6b9-2bd8553a6ed3 service nova] Acquired lock "refresh_cache-13d20f82-fa4f-41ab-b08f-5b64da7dddc0" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.804879] env[62965]: DEBUG nova.network.neutron [req-93a928bd-c56e-4b0e-a644-02c4938fd7f2 req-f838dd22-f0a8-484b-a6b9-2bd8553a6ed3 service nova] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Refreshing network info cache for port 6d23ec55-580a-4388-b3ad-9bfa3d9b47bb {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 818.805960] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-45909ac4-f72e-49ce-b075-2db1070fb250 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.816647] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcd37a60-4151-4ad4-b455-5b27628fcb67 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.840677] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 13d20f82-fa4f-41ab-b08f-5b64da7dddc0 could not be found. [ 818.840890] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 818.841083] env[62965]: INFO nova.compute.manager [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Took 0.04 seconds to destroy the instance on the hypervisor. [ 818.841328] env[62965]: DEBUG oslo.service.loopingcall [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 818.841580] env[62965]: DEBUG nova.compute.manager [-] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 818.841661] env[62965]: DEBUG nova.network.neutron [-] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 818.869009] env[62965]: DEBUG nova.network.neutron [-] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 818.902296] env[62965]: DEBUG nova.compute.utils [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 818.904634] env[62965]: DEBUG nova.compute.manager [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 818.904878] env[62965]: DEBUG nova.network.neutron [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 818.944170] env[62965]: DEBUG nova.policy [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1436dc87ceab467fb4d95604dd774da0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '79c5dcb49c454d85a8588e3d89be83e8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 819.125910] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26870ed9-8e37-49f8-8af6-5e2ecada95ec {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.133310] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eec119d-def9-4f9a-83b6-68d5714fdfe1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.162902] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8222908f-b883-4ed6-8d5d-390bf5a45b3e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.171740] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-977020ad-a6c7-4658-b1e5-327fb1953e0f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.192113] env[62965]: DEBUG nova.compute.provider_tree [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 819.241113] env[62965]: DEBUG nova.network.neutron [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Successfully created port: c868142e-9eff-4886-a48b-aca62edbd872 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 819.328821] env[62965]: DEBUG nova.network.neutron [req-93a928bd-c56e-4b0e-a644-02c4938fd7f2 req-f838dd22-f0a8-484b-a6b9-2bd8553a6ed3 service nova] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 819.371231] env[62965]: DEBUG nova.network.neutron [-] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.405478] env[62965]: DEBUG nova.compute.manager [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 819.421944] env[62965]: DEBUG nova.network.neutron [req-93a928bd-c56e-4b0e-a644-02c4938fd7f2 req-f838dd22-f0a8-484b-a6b9-2bd8553a6ed3 service nova] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.689711] env[62965]: DEBUG nova.scheduler.client.report [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 819.875029] env[62965]: INFO nova.compute.manager [-] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Took 1.03 seconds to deallocate network for instance. [ 819.876757] env[62965]: DEBUG nova.compute.claims [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 819.876965] env[62965]: DEBUG oslo_concurrency.lockutils [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.924581] env[62965]: DEBUG oslo_concurrency.lockutils [req-93a928bd-c56e-4b0e-a644-02c4938fd7f2 req-f838dd22-f0a8-484b-a6b9-2bd8553a6ed3 service nova] Releasing lock "refresh_cache-13d20f82-fa4f-41ab-b08f-5b64da7dddc0" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.924986] env[62965]: DEBUG nova.compute.manager [req-93a928bd-c56e-4b0e-a644-02c4938fd7f2 req-f838dd22-f0a8-484b-a6b9-2bd8553a6ed3 service nova] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Received event network-vif-deleted-6d23ec55-580a-4388-b3ad-9bfa3d9b47bb {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 820.187139] env[62965]: DEBUG nova.compute.manager [req-80b67a84-751d-4a94-adf4-e2a7c7c07973 req-4669d94c-2c14-4e39-a9e2-12daecab5440 service nova] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Received event network-changed-c868142e-9eff-4886-a48b-aca62edbd872 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 820.187447] env[62965]: DEBUG nova.compute.manager [req-80b67a84-751d-4a94-adf4-e2a7c7c07973 req-4669d94c-2c14-4e39-a9e2-12daecab5440 service nova] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Refreshing instance network info cache due to event network-changed-c868142e-9eff-4886-a48b-aca62edbd872. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 820.187732] env[62965]: DEBUG oslo_concurrency.lockutils [req-80b67a84-751d-4a94-adf4-e2a7c7c07973 req-4669d94c-2c14-4e39-a9e2-12daecab5440 service nova] Acquiring lock "refresh_cache-474c8faf-cf69-4296-8fba-521e5472c071" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.188027] env[62965]: DEBUG oslo_concurrency.lockutils [req-80b67a84-751d-4a94-adf4-e2a7c7c07973 req-4669d94c-2c14-4e39-a9e2-12daecab5440 service nova] Acquired lock "refresh_cache-474c8faf-cf69-4296-8fba-521e5472c071" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.188231] env[62965]: DEBUG nova.network.neutron [req-80b67a84-751d-4a94-adf4-e2a7c7c07973 req-4669d94c-2c14-4e39-a9e2-12daecab5440 service nova] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Refreshing network info cache for port c868142e-9eff-4886-a48b-aca62edbd872 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 820.197987] env[62965]: DEBUG oslo_concurrency.lockutils [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.801s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.200506] env[62965]: DEBUG oslo_concurrency.lockutils [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.275s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.201914] env[62965]: INFO nova.compute.claims [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 820.222796] env[62965]: INFO nova.scheduler.client.report [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Deleted allocations for instance e87adfae-3ea4-4cc4-9978-d8c811c80941 [ 820.305752] env[62965]: ERROR nova.compute.manager [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c868142e-9eff-4886-a48b-aca62edbd872, please check neutron logs for more information. [ 820.305752] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 820.305752] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 820.305752] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 820.305752] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 820.305752] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 820.305752] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 820.305752] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 820.305752] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 820.305752] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 820.305752] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 820.305752] env[62965]: ERROR nova.compute.manager raise self.value [ 820.305752] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 820.305752] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 820.305752] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 820.305752] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 820.306179] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 820.306179] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 820.306179] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c868142e-9eff-4886-a48b-aca62edbd872, please check neutron logs for more information. [ 820.306179] env[62965]: ERROR nova.compute.manager [ 820.306179] env[62965]: Traceback (most recent call last): [ 820.306179] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 820.306179] env[62965]: listener.cb(fileno) [ 820.306179] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 820.306179] env[62965]: result = function(*args, **kwargs) [ 820.306179] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 820.306179] env[62965]: return func(*args, **kwargs) [ 820.306179] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 820.306179] env[62965]: raise e [ 820.306179] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 820.306179] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 820.306179] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 820.306179] env[62965]: created_port_ids = self._update_ports_for_instance( [ 820.306179] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 820.306179] env[62965]: with excutils.save_and_reraise_exception(): [ 820.306179] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 820.306179] env[62965]: self.force_reraise() [ 820.306179] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 820.306179] env[62965]: raise self.value [ 820.306179] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 820.306179] env[62965]: updated_port = self._update_port( [ 820.306179] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 820.306179] env[62965]: _ensure_no_port_binding_failure(port) [ 820.306179] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 820.306179] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 820.306872] env[62965]: nova.exception.PortBindingFailed: Binding failed for port c868142e-9eff-4886-a48b-aca62edbd872, please check neutron logs for more information. [ 820.306872] env[62965]: Removing descriptor: 17 [ 820.414690] env[62965]: DEBUG nova.compute.manager [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 820.439759] env[62965]: DEBUG nova.virt.hardware [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 820.439998] env[62965]: DEBUG nova.virt.hardware [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 820.440161] env[62965]: DEBUG nova.virt.hardware [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 820.440373] env[62965]: DEBUG nova.virt.hardware [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 820.440524] env[62965]: DEBUG nova.virt.hardware [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 820.440667] env[62965]: DEBUG nova.virt.hardware [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 820.440869] env[62965]: DEBUG nova.virt.hardware [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 820.441036] env[62965]: DEBUG nova.virt.hardware [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 820.441207] env[62965]: DEBUG nova.virt.hardware [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 820.441366] env[62965]: DEBUG nova.virt.hardware [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 820.441589] env[62965]: DEBUG nova.virt.hardware [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 820.442850] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43c8049a-b3d0-4f76-bfef-331773e17c8f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.450740] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bda28be4-f857-4226-a9ff-1f931b084c99 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.464243] env[62965]: ERROR nova.compute.manager [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c868142e-9eff-4886-a48b-aca62edbd872, please check neutron logs for more information. [ 820.464243] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Traceback (most recent call last): [ 820.464243] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 820.464243] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] yield resources [ 820.464243] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 820.464243] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] self.driver.spawn(context, instance, image_meta, [ 820.464243] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 820.464243] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] self._vmops.spawn(context, instance, image_meta, injected_files, [ 820.464243] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 820.464243] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] vm_ref = self.build_virtual_machine(instance, [ 820.464243] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 820.464576] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] vif_infos = vmwarevif.get_vif_info(self._session, [ 820.464576] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 820.464576] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] for vif in network_info: [ 820.464576] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 820.464576] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] return self._sync_wrapper(fn, *args, **kwargs) [ 820.464576] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 820.464576] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] self.wait() [ 820.464576] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 820.464576] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] self[:] = self._gt.wait() [ 820.464576] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 820.464576] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] return self._exit_event.wait() [ 820.464576] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 820.464576] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] current.throw(*self._exc) [ 820.464927] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 820.464927] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] result = function(*args, **kwargs) [ 820.464927] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 820.464927] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] return func(*args, **kwargs) [ 820.464927] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 820.464927] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] raise e [ 820.464927] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 820.464927] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] nwinfo = self.network_api.allocate_for_instance( [ 820.464927] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 820.464927] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] created_port_ids = self._update_ports_for_instance( [ 820.464927] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 820.464927] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] with excutils.save_and_reraise_exception(): [ 820.464927] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 820.465281] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] self.force_reraise() [ 820.465281] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 820.465281] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] raise self.value [ 820.465281] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 820.465281] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] updated_port = self._update_port( [ 820.465281] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 820.465281] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] _ensure_no_port_binding_failure(port) [ 820.465281] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 820.465281] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] raise exception.PortBindingFailed(port_id=port['id']) [ 820.465281] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] nova.exception.PortBindingFailed: Binding failed for port c868142e-9eff-4886-a48b-aca62edbd872, please check neutron logs for more information. [ 820.465281] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] [ 820.465281] env[62965]: INFO nova.compute.manager [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Terminating instance [ 820.708143] env[62965]: DEBUG nova.network.neutron [req-80b67a84-751d-4a94-adf4-e2a7c7c07973 req-4669d94c-2c14-4e39-a9e2-12daecab5440 service nova] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 820.730034] env[62965]: DEBUG oslo_concurrency.lockutils [None req-275fd008-c14e-4617-815d-c6f306856afd tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Lock "e87adfae-3ea4-4cc4-9978-d8c811c80941" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.233s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.797398] env[62965]: DEBUG nova.network.neutron [req-80b67a84-751d-4a94-adf4-e2a7c7c07973 req-4669d94c-2c14-4e39-a9e2-12daecab5440 service nova] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.971166] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] Acquiring lock "refresh_cache-474c8faf-cf69-4296-8fba-521e5472c071" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.971569] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Acquiring lock "6e0a4e4b-21a2-4185-999e-69b4c74c37d7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.971835] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Lock "6e0a4e4b-21a2-4185-999e-69b4c74c37d7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.972082] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Acquiring lock "6e0a4e4b-21a2-4185-999e-69b4c74c37d7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.972314] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Lock "6e0a4e4b-21a2-4185-999e-69b4c74c37d7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.972520] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Lock "6e0a4e4b-21a2-4185-999e-69b4c74c37d7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.974689] env[62965]: INFO nova.compute.manager [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Terminating instance [ 821.301663] env[62965]: DEBUG oslo_concurrency.lockutils [req-80b67a84-751d-4a94-adf4-e2a7c7c07973 req-4669d94c-2c14-4e39-a9e2-12daecab5440 service nova] Releasing lock "refresh_cache-474c8faf-cf69-4296-8fba-521e5472c071" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.301954] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] Acquired lock "refresh_cache-474c8faf-cf69-4296-8fba-521e5472c071" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.302162] env[62965]: DEBUG nova.network.neutron [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 821.418071] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80a7cf46-00c5-45f9-9f43-531a62beaa41 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.427102] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c2c6f30-1ce7-4cf5-bee0-e38bee29a1c0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.456624] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14252f41-4b87-4131-aa3f-e8a7e544a268 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.463483] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57363f4e-e01b-4b3a-a23c-c6ac1f40c36b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.476157] env[62965]: DEBUG nova.compute.provider_tree [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 821.477629] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Acquiring lock "refresh_cache-6e0a4e4b-21a2-4185-999e-69b4c74c37d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.477784] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Acquired lock "refresh_cache-6e0a4e4b-21a2-4185-999e-69b4c74c37d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.477949] env[62965]: DEBUG nova.network.neutron [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 821.820446] env[62965]: DEBUG nova.network.neutron [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 821.884529] env[62965]: DEBUG nova.network.neutron [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.980151] env[62965]: DEBUG nova.scheduler.client.report [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 822.016577] env[62965]: DEBUG nova.network.neutron [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 822.083177] env[62965]: DEBUG nova.network.neutron [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.213067] env[62965]: DEBUG nova.compute.manager [req-23b5e5ce-1f08-4097-9edf-7992ba901927 req-799119b1-6623-4408-84de-0d433ad9349d service nova] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Received event network-vif-deleted-c868142e-9eff-4886-a48b-aca62edbd872 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 822.386996] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] Releasing lock "refresh_cache-474c8faf-cf69-4296-8fba-521e5472c071" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.387483] env[62965]: DEBUG nova.compute.manager [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 822.387677] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 822.387981] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c8d1ea7d-c77a-430a-adc4-1d17e26b7c4f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.397137] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e68b29e-0f6c-45ad-98cc-53d0c0accf17 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.420390] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 474c8faf-cf69-4296-8fba-521e5472c071 could not be found. [ 822.420611] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 822.420812] env[62965]: INFO nova.compute.manager [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Took 0.03 seconds to destroy the instance on the hypervisor. [ 822.421065] env[62965]: DEBUG oslo.service.loopingcall [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 822.421290] env[62965]: DEBUG nova.compute.manager [-] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 822.421387] env[62965]: DEBUG nova.network.neutron [-] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 822.434297] env[62965]: DEBUG nova.network.neutron [-] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 822.486594] env[62965]: DEBUG oslo_concurrency.lockutils [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.286s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.487192] env[62965]: DEBUG nova.compute.manager [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 822.489676] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.729s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.491506] env[62965]: INFO nova.compute.claims [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 822.585294] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Releasing lock "refresh_cache-6e0a4e4b-21a2-4185-999e-69b4c74c37d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.585716] env[62965]: DEBUG nova.compute.manager [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 822.585904] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 822.586784] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dba09ffd-b6a1-41c0-a91e-8be7bd81695a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.595557] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 822.595834] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ce030725-7a4c-4df4-8298-1b1aca74f7fb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.601831] env[62965]: DEBUG oslo_vmware.api [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Waiting for the task: (returnval){ [ 822.601831] env[62965]: value = "task-1389811" [ 822.601831] env[62965]: _type = "Task" [ 822.601831] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.609381] env[62965]: DEBUG oslo_vmware.api [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389811, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.939062] env[62965]: DEBUG nova.network.neutron [-] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.995913] env[62965]: DEBUG nova.compute.utils [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 822.999838] env[62965]: DEBUG nova.compute.manager [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 822.999997] env[62965]: DEBUG nova.network.neutron [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 823.049104] env[62965]: DEBUG nova.policy [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '796be5c70f074929943bc1196791eba3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '07cc096d8b1e4f8284d9349b038e430e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 823.111103] env[62965]: DEBUG oslo_vmware.api [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389811, 'name': PowerOffVM_Task, 'duration_secs': 0.151689} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.111380] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 823.111547] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 823.111790] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f5201bcc-01aa-468d-8487-862af33cae1c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.136211] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 823.136431] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Deleting contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 823.136613] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Deleting the datastore file [datastore2] 6e0a4e4b-21a2-4185-999e-69b4c74c37d7 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 823.136867] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8096174a-13ae-4b4b-8c50-09fe74ca45f2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.143245] env[62965]: DEBUG oslo_vmware.api [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Waiting for the task: (returnval){ [ 823.143245] env[62965]: value = "task-1389813" [ 823.143245] env[62965]: _type = "Task" [ 823.143245] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.150771] env[62965]: DEBUG oslo_vmware.api [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389813, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.370510] env[62965]: DEBUG nova.network.neutron [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Successfully created port: f63f2a62-43a8-4fdb-9668-0dd96a228033 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 823.442160] env[62965]: INFO nova.compute.manager [-] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Took 1.02 seconds to deallocate network for instance. [ 823.444694] env[62965]: DEBUG nova.compute.claims [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 823.444872] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.500781] env[62965]: DEBUG nova.compute.manager [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 823.660096] env[62965]: DEBUG oslo_vmware.api [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Task: {'id': task-1389813, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.085591} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.660096] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 823.660096] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Deleted contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 823.660096] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 823.660096] env[62965]: INFO nova.compute.manager [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Took 1.07 seconds to destroy the instance on the hypervisor. [ 823.660391] env[62965]: DEBUG oslo.service.loopingcall [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 823.660391] env[62965]: DEBUG nova.compute.manager [-] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 823.660391] env[62965]: DEBUG nova.network.neutron [-] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 823.675627] env[62965]: DEBUG nova.network.neutron [-] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 823.745833] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b0ebb95-fef3-47cd-8de2-196d2e09124b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.753295] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4d1d405-fbbb-4d2d-a5c6-5002ab0d9197 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.783803] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-781ccf42-3a84-4041-b35c-297b6ed45227 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.791090] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e308371-5c16-4710-b264-519ed099e5cb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.804016] env[62965]: DEBUG nova.compute.provider_tree [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 824.177924] env[62965]: DEBUG nova.network.neutron [-] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.307310] env[62965]: DEBUG nova.scheduler.client.report [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 824.395585] env[62965]: DEBUG nova.compute.manager [req-d8b12cdd-3c81-4b16-b73b-cbb11c534a20 req-223e06d5-ef62-456b-9e07-e2fcfaf34fd4 service nova] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Received event network-changed-f63f2a62-43a8-4fdb-9668-0dd96a228033 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 824.395791] env[62965]: DEBUG nova.compute.manager [req-d8b12cdd-3c81-4b16-b73b-cbb11c534a20 req-223e06d5-ef62-456b-9e07-e2fcfaf34fd4 service nova] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Refreshing instance network info cache due to event network-changed-f63f2a62-43a8-4fdb-9668-0dd96a228033. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 824.396012] env[62965]: DEBUG oslo_concurrency.lockutils [req-d8b12cdd-3c81-4b16-b73b-cbb11c534a20 req-223e06d5-ef62-456b-9e07-e2fcfaf34fd4 service nova] Acquiring lock "refresh_cache-e9cac9cf-c165-428b-9d83-e37905f42c93" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.396219] env[62965]: DEBUG oslo_concurrency.lockutils [req-d8b12cdd-3c81-4b16-b73b-cbb11c534a20 req-223e06d5-ef62-456b-9e07-e2fcfaf34fd4 service nova] Acquired lock "refresh_cache-e9cac9cf-c165-428b-9d83-e37905f42c93" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.396405] env[62965]: DEBUG nova.network.neutron [req-d8b12cdd-3c81-4b16-b73b-cbb11c534a20 req-223e06d5-ef62-456b-9e07-e2fcfaf34fd4 service nova] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Refreshing network info cache for port f63f2a62-43a8-4fdb-9668-0dd96a228033 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 824.511988] env[62965]: DEBUG nova.compute.manager [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 824.536033] env[62965]: DEBUG nova.virt.hardware [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:34:20Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='c28b0297-a5cc-423f-8177-c82aebb0d8ef',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-274482305',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 824.536206] env[62965]: DEBUG nova.virt.hardware [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 824.537020] env[62965]: DEBUG nova.virt.hardware [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 824.537020] env[62965]: DEBUG nova.virt.hardware [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 824.537020] env[62965]: DEBUG nova.virt.hardware [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 824.537020] env[62965]: DEBUG nova.virt.hardware [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 824.537020] env[62965]: DEBUG nova.virt.hardware [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 824.537241] env[62965]: DEBUG nova.virt.hardware [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 824.537319] env[62965]: DEBUG nova.virt.hardware [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 824.537476] env[62965]: DEBUG nova.virt.hardware [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 824.537644] env[62965]: DEBUG nova.virt.hardware [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 824.538826] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20f7d6ab-759d-4bbd-bdae-5f6e8bd83bbf {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.546541] env[62965]: ERROR nova.compute.manager [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f63f2a62-43a8-4fdb-9668-0dd96a228033, please check neutron logs for more information. [ 824.546541] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 824.546541] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 824.546541] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 824.546541] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 824.546541] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 824.546541] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 824.546541] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 824.546541] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 824.546541] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 824.546541] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 824.546541] env[62965]: ERROR nova.compute.manager raise self.value [ 824.546541] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 824.546541] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 824.546541] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 824.546541] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 824.546974] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 824.546974] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 824.546974] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f63f2a62-43a8-4fdb-9668-0dd96a228033, please check neutron logs for more information. [ 824.546974] env[62965]: ERROR nova.compute.manager [ 824.546974] env[62965]: Traceback (most recent call last): [ 824.546974] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 824.546974] env[62965]: listener.cb(fileno) [ 824.546974] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 824.546974] env[62965]: result = function(*args, **kwargs) [ 824.546974] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 824.546974] env[62965]: return func(*args, **kwargs) [ 824.546974] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 824.546974] env[62965]: raise e [ 824.546974] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 824.546974] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 824.546974] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 824.546974] env[62965]: created_port_ids = self._update_ports_for_instance( [ 824.546974] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 824.546974] env[62965]: with excutils.save_and_reraise_exception(): [ 824.546974] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 824.546974] env[62965]: self.force_reraise() [ 824.546974] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 824.546974] env[62965]: raise self.value [ 824.546974] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 824.546974] env[62965]: updated_port = self._update_port( [ 824.546974] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 824.546974] env[62965]: _ensure_no_port_binding_failure(port) [ 824.546974] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 824.546974] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 824.547738] env[62965]: nova.exception.PortBindingFailed: Binding failed for port f63f2a62-43a8-4fdb-9668-0dd96a228033, please check neutron logs for more information. [ 824.547738] env[62965]: Removing descriptor: 17 [ 824.547910] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80b66296-2cec-4c1d-b7d3-d91f08f3c975 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.562278] env[62965]: ERROR nova.compute.manager [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f63f2a62-43a8-4fdb-9668-0dd96a228033, please check neutron logs for more information. [ 824.562278] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Traceback (most recent call last): [ 824.562278] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 824.562278] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] yield resources [ 824.562278] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 824.562278] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] self.driver.spawn(context, instance, image_meta, [ 824.562278] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 824.562278] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] self._vmops.spawn(context, instance, image_meta, injected_files, [ 824.562278] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 824.562278] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] vm_ref = self.build_virtual_machine(instance, [ 824.562278] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 824.562760] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] vif_infos = vmwarevif.get_vif_info(self._session, [ 824.562760] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 824.562760] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] for vif in network_info: [ 824.562760] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 824.562760] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] return self._sync_wrapper(fn, *args, **kwargs) [ 824.562760] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 824.562760] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] self.wait() [ 824.562760] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 824.562760] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] self[:] = self._gt.wait() [ 824.562760] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 824.562760] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] return self._exit_event.wait() [ 824.562760] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 824.562760] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] current.throw(*self._exc) [ 824.563111] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 824.563111] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] result = function(*args, **kwargs) [ 824.563111] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 824.563111] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] return func(*args, **kwargs) [ 824.563111] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 824.563111] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] raise e [ 824.563111] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 824.563111] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] nwinfo = self.network_api.allocate_for_instance( [ 824.563111] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 824.563111] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] created_port_ids = self._update_ports_for_instance( [ 824.563111] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 824.563111] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] with excutils.save_and_reraise_exception(): [ 824.563111] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 824.563452] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] self.force_reraise() [ 824.563452] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 824.563452] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] raise self.value [ 824.563452] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 824.563452] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] updated_port = self._update_port( [ 824.563452] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 824.563452] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] _ensure_no_port_binding_failure(port) [ 824.563452] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 824.563452] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] raise exception.PortBindingFailed(port_id=port['id']) [ 824.563452] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] nova.exception.PortBindingFailed: Binding failed for port f63f2a62-43a8-4fdb-9668-0dd96a228033, please check neutron logs for more information. [ 824.563452] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] [ 824.563452] env[62965]: INFO nova.compute.manager [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Terminating instance [ 824.680196] env[62965]: INFO nova.compute.manager [-] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Took 1.02 seconds to deallocate network for instance. [ 824.812337] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.322s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.812866] env[62965]: DEBUG nova.compute.manager [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 824.815580] env[62965]: DEBUG oslo_concurrency.lockutils [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.340s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.912723] env[62965]: DEBUG nova.network.neutron [req-d8b12cdd-3c81-4b16-b73b-cbb11c534a20 req-223e06d5-ef62-456b-9e07-e2fcfaf34fd4 service nova] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 824.971692] env[62965]: DEBUG nova.network.neutron [req-d8b12cdd-3c81-4b16-b73b-cbb11c534a20 req-223e06d5-ef62-456b-9e07-e2fcfaf34fd4 service nova] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.066355] env[62965]: DEBUG oslo_concurrency.lockutils [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Acquiring lock "refresh_cache-e9cac9cf-c165-428b-9d83-e37905f42c93" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.187228] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.320653] env[62965]: DEBUG nova.compute.utils [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 825.325571] env[62965]: DEBUG nova.compute.manager [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 825.325743] env[62965]: DEBUG nova.network.neutron [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 825.374754] env[62965]: DEBUG nova.policy [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd4224e7cf4154d23bbeb09d7d9301393', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7b91fa28e13d4212b91014f553f866a6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 825.473649] env[62965]: DEBUG oslo_concurrency.lockutils [req-d8b12cdd-3c81-4b16-b73b-cbb11c534a20 req-223e06d5-ef62-456b-9e07-e2fcfaf34fd4 service nova] Releasing lock "refresh_cache-e9cac9cf-c165-428b-9d83-e37905f42c93" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.474070] env[62965]: DEBUG oslo_concurrency.lockutils [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Acquired lock "refresh_cache-e9cac9cf-c165-428b-9d83-e37905f42c93" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.474228] env[62965]: DEBUG nova.network.neutron [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 825.556039] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-034b77ce-022f-4670-bcd0-78368f3f1f2a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.563826] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-235de932-1fbb-4ac3-8f45-89f8b813a42e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.594255] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f88c0bb9-94d1-4373-9efa-5d53c0b19cba {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.601405] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b9b9158-ddbd-491a-879b-a875721a744c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.615788] env[62965]: DEBUG nova.compute.provider_tree [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 825.680227] env[62965]: DEBUG nova.network.neutron [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Successfully created port: e76c7578-9d13-4705-8a04-501e48bed907 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 825.826108] env[62965]: DEBUG nova.compute.manager [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 825.999165] env[62965]: DEBUG nova.network.neutron [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 826.049934] env[62965]: DEBUG nova.network.neutron [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.122405] env[62965]: DEBUG nova.scheduler.client.report [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 826.421987] env[62965]: DEBUG nova.compute.manager [req-4d9dce18-f104-424b-a1e8-4687d3f3bab3 req-15c7b77d-c71b-4161-94a8-3a97f187c933 service nova] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Received event network-vif-deleted-f63f2a62-43a8-4fdb-9668-0dd96a228033 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 826.555888] env[62965]: DEBUG oslo_concurrency.lockutils [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Releasing lock "refresh_cache-e9cac9cf-c165-428b-9d83-e37905f42c93" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.556308] env[62965]: DEBUG nova.compute.manager [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 826.556558] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 826.556776] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e833eae8-e25d-4c06-8411-7dcc613341ae {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.566256] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-619f20fa-ecad-455f-b270-2aafc01a1663 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.588570] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e9cac9cf-c165-428b-9d83-e37905f42c93 could not be found. [ 826.588570] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 826.588570] env[62965]: INFO nova.compute.manager [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Took 0.03 seconds to destroy the instance on the hypervisor. [ 826.588570] env[62965]: DEBUG oslo.service.loopingcall [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 826.588771] env[62965]: DEBUG nova.compute.manager [-] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 826.588771] env[62965]: DEBUG nova.network.neutron [-] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 826.604922] env[62965]: DEBUG nova.network.neutron [-] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 826.624352] env[62965]: DEBUG oslo_concurrency.lockutils [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.809s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.624940] env[62965]: ERROR nova.compute.manager [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ac88ed76-44e1-46ea-9dc4-31e693ea44a6, please check neutron logs for more information. [ 826.624940] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Traceback (most recent call last): [ 826.624940] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 826.624940] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] self.driver.spawn(context, instance, image_meta, [ 826.624940] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 826.624940] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 826.624940] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 826.624940] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] vm_ref = self.build_virtual_machine(instance, [ 826.624940] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 826.624940] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] vif_infos = vmwarevif.get_vif_info(self._session, [ 826.624940] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 826.625261] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] for vif in network_info: [ 826.625261] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 826.625261] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] return self._sync_wrapper(fn, *args, **kwargs) [ 826.625261] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 826.625261] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] self.wait() [ 826.625261] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 826.625261] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] self[:] = self._gt.wait() [ 826.625261] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 826.625261] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] return self._exit_event.wait() [ 826.625261] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 826.625261] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] result = hub.switch() [ 826.625261] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 826.625261] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] return self.greenlet.switch() [ 826.625607] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 826.625607] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] result = function(*args, **kwargs) [ 826.625607] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 826.625607] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] return func(*args, **kwargs) [ 826.625607] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 826.625607] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] raise e [ 826.625607] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 826.625607] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] nwinfo = self.network_api.allocate_for_instance( [ 826.625607] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 826.625607] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] created_port_ids = self._update_ports_for_instance( [ 826.625607] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 826.625607] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] with excutils.save_and_reraise_exception(): [ 826.625607] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 826.626016] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] self.force_reraise() [ 826.626016] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 826.626016] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] raise self.value [ 826.626016] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 826.626016] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] updated_port = self._update_port( [ 826.626016] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 826.626016] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] _ensure_no_port_binding_failure(port) [ 826.626016] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 826.626016] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] raise exception.PortBindingFailed(port_id=port['id']) [ 826.626016] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] nova.exception.PortBindingFailed: Binding failed for port ac88ed76-44e1-46ea-9dc4-31e693ea44a6, please check neutron logs for more information. [ 826.626016] env[62965]: ERROR nova.compute.manager [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] [ 826.626315] env[62965]: DEBUG nova.compute.utils [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Binding failed for port ac88ed76-44e1-46ea-9dc4-31e693ea44a6, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 826.626730] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.416s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.629558] env[62965]: DEBUG nova.compute.manager [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Build of instance 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1 was re-scheduled: Binding failed for port ac88ed76-44e1-46ea-9dc4-31e693ea44a6, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 826.630011] env[62965]: DEBUG nova.compute.manager [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 826.630540] env[62965]: DEBUG oslo_concurrency.lockutils [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] Acquiring lock "refresh_cache-6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.630702] env[62965]: DEBUG oslo_concurrency.lockutils [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] Acquired lock "refresh_cache-6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.630862] env[62965]: DEBUG nova.network.neutron [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 826.697401] env[62965]: ERROR nova.compute.manager [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e76c7578-9d13-4705-8a04-501e48bed907, please check neutron logs for more information. [ 826.697401] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 826.697401] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 826.697401] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 826.697401] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 826.697401] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 826.697401] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 826.697401] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 826.697401] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 826.697401] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 826.697401] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 826.697401] env[62965]: ERROR nova.compute.manager raise self.value [ 826.697401] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 826.697401] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 826.697401] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 826.697401] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 826.697827] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 826.697827] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 826.697827] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e76c7578-9d13-4705-8a04-501e48bed907, please check neutron logs for more information. [ 826.697827] env[62965]: ERROR nova.compute.manager [ 826.697827] env[62965]: Traceback (most recent call last): [ 826.697827] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 826.697827] env[62965]: listener.cb(fileno) [ 826.697827] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 826.697827] env[62965]: result = function(*args, **kwargs) [ 826.697827] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 826.697827] env[62965]: return func(*args, **kwargs) [ 826.697827] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 826.697827] env[62965]: raise e [ 826.697827] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 826.697827] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 826.697827] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 826.697827] env[62965]: created_port_ids = self._update_ports_for_instance( [ 826.697827] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 826.697827] env[62965]: with excutils.save_and_reraise_exception(): [ 826.697827] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 826.697827] env[62965]: self.force_reraise() [ 826.697827] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 826.697827] env[62965]: raise self.value [ 826.697827] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 826.697827] env[62965]: updated_port = self._update_port( [ 826.697827] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 826.697827] env[62965]: _ensure_no_port_binding_failure(port) [ 826.697827] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 826.697827] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 826.698597] env[62965]: nova.exception.PortBindingFailed: Binding failed for port e76c7578-9d13-4705-8a04-501e48bed907, please check neutron logs for more information. [ 826.698597] env[62965]: Removing descriptor: 17 [ 826.837184] env[62965]: DEBUG nova.compute.manager [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 826.861127] env[62965]: DEBUG nova.virt.hardware [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 826.861127] env[62965]: DEBUG nova.virt.hardware [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 826.861127] env[62965]: DEBUG nova.virt.hardware [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 826.861573] env[62965]: DEBUG nova.virt.hardware [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 826.861832] env[62965]: DEBUG nova.virt.hardware [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 826.862109] env[62965]: DEBUG nova.virt.hardware [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 826.862423] env[62965]: DEBUG nova.virt.hardware [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 826.862690] env[62965]: DEBUG nova.virt.hardware [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 826.862983] env[62965]: DEBUG nova.virt.hardware [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 826.864740] env[62965]: DEBUG nova.virt.hardware [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 826.864740] env[62965]: DEBUG nova.virt.hardware [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 826.864740] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c759a78-ea2c-4101-b4dc-ddf2e92c603e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.872184] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bb40935-a829-412e-8cbc-ea80e8607fac {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.886868] env[62965]: ERROR nova.compute.manager [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e76c7578-9d13-4705-8a04-501e48bed907, please check neutron logs for more information. [ 826.886868] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Traceback (most recent call last): [ 826.886868] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 826.886868] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] yield resources [ 826.886868] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 826.886868] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] self.driver.spawn(context, instance, image_meta, [ 826.886868] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 826.886868] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 826.886868] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 826.886868] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] vm_ref = self.build_virtual_machine(instance, [ 826.886868] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 826.887237] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] vif_infos = vmwarevif.get_vif_info(self._session, [ 826.887237] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 826.887237] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] for vif in network_info: [ 826.887237] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 826.887237] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] return self._sync_wrapper(fn, *args, **kwargs) [ 826.887237] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 826.887237] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] self.wait() [ 826.887237] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 826.887237] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] self[:] = self._gt.wait() [ 826.887237] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 826.887237] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] return self._exit_event.wait() [ 826.887237] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 826.887237] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] current.throw(*self._exc) [ 826.887656] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 826.887656] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] result = function(*args, **kwargs) [ 826.887656] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 826.887656] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] return func(*args, **kwargs) [ 826.887656] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 826.887656] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] raise e [ 826.887656] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 826.887656] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] nwinfo = self.network_api.allocate_for_instance( [ 826.887656] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 826.887656] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] created_port_ids = self._update_ports_for_instance( [ 826.887656] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 826.887656] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] with excutils.save_and_reraise_exception(): [ 826.887656] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 826.888057] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] self.force_reraise() [ 826.888057] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 826.888057] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] raise self.value [ 826.888057] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 826.888057] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] updated_port = self._update_port( [ 826.888057] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 826.888057] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] _ensure_no_port_binding_failure(port) [ 826.888057] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 826.888057] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] raise exception.PortBindingFailed(port_id=port['id']) [ 826.888057] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] nova.exception.PortBindingFailed: Binding failed for port e76c7578-9d13-4705-8a04-501e48bed907, please check neutron logs for more information. [ 826.888057] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] [ 826.888057] env[62965]: INFO nova.compute.manager [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Terminating instance [ 827.107492] env[62965]: DEBUG nova.network.neutron [-] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.158031] env[62965]: DEBUG nova.network.neutron [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 827.250775] env[62965]: DEBUG nova.network.neutron [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.369668] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-378dcd8f-3625-4102-a2c3-eb83acc20d4f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.377729] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fce10bc-8d17-4683-9a70-ba430057badb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.406387] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] Acquiring lock "refresh_cache-3d0f74d4-ff59-462c-9577-c01c4f15ea1d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.406563] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] Acquired lock "refresh_cache-3d0f74d4-ff59-462c-9577-c01c4f15ea1d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.406735] env[62965]: DEBUG nova.network.neutron [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 827.408638] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a69224b-074e-4e81-bf12-996be614f150 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.415540] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b02e442-3aca-46ee-aeea-a23f685e0c0d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.429041] env[62965]: DEBUG nova.compute.provider_tree [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 827.610105] env[62965]: INFO nova.compute.manager [-] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Took 1.02 seconds to deallocate network for instance. [ 827.612086] env[62965]: DEBUG nova.compute.claims [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 827.612265] env[62965]: DEBUG oslo_concurrency.lockutils [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.756626] env[62965]: DEBUG oslo_concurrency.lockutils [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] Releasing lock "refresh_cache-6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.756872] env[62965]: DEBUG nova.compute.manager [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 827.757075] env[62965]: DEBUG nova.compute.manager [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 827.757244] env[62965]: DEBUG nova.network.neutron [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 827.772019] env[62965]: DEBUG nova.network.neutron [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 827.932180] env[62965]: DEBUG nova.scheduler.client.report [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 827.941418] env[62965]: DEBUG nova.network.neutron [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 828.054703] env[62965]: DEBUG nova.network.neutron [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.275169] env[62965]: DEBUG nova.network.neutron [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.437111] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.810s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.437865] env[62965]: ERROR nova.compute.manager [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 71c05380-b654-4e8a-addc-e438efd4e9ca, please check neutron logs for more information. [ 828.437865] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Traceback (most recent call last): [ 828.437865] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 828.437865] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] self.driver.spawn(context, instance, image_meta, [ 828.437865] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 828.437865] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 828.437865] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 828.437865] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] vm_ref = self.build_virtual_machine(instance, [ 828.437865] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 828.437865] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] vif_infos = vmwarevif.get_vif_info(self._session, [ 828.437865] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 828.438269] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] for vif in network_info: [ 828.438269] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 828.438269] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] return self._sync_wrapper(fn, *args, **kwargs) [ 828.438269] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 828.438269] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] self.wait() [ 828.438269] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 828.438269] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] self[:] = self._gt.wait() [ 828.438269] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 828.438269] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] return self._exit_event.wait() [ 828.438269] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 828.438269] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] current.throw(*self._exc) [ 828.438269] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 828.438269] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] result = function(*args, **kwargs) [ 828.438621] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 828.438621] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] return func(*args, **kwargs) [ 828.438621] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 828.438621] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] raise e [ 828.438621] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 828.438621] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] nwinfo = self.network_api.allocate_for_instance( [ 828.438621] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 828.438621] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] created_port_ids = self._update_ports_for_instance( [ 828.438621] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 828.438621] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] with excutils.save_and_reraise_exception(): [ 828.438621] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 828.438621] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] self.force_reraise() [ 828.438621] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 828.438988] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] raise self.value [ 828.438988] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 828.438988] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] updated_port = self._update_port( [ 828.438988] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 828.438988] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] _ensure_no_port_binding_failure(port) [ 828.438988] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 828.438988] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] raise exception.PortBindingFailed(port_id=port['id']) [ 828.438988] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] nova.exception.PortBindingFailed: Binding failed for port 71c05380-b654-4e8a-addc-e438efd4e9ca, please check neutron logs for more information. [ 828.438988] env[62965]: ERROR nova.compute.manager [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] [ 828.438988] env[62965]: DEBUG nova.compute.utils [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Binding failed for port 71c05380-b654-4e8a-addc-e438efd4e9ca, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 828.439344] env[62965]: DEBUG oslo_concurrency.lockutils [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.735s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.441163] env[62965]: INFO nova.compute.claims [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 828.443917] env[62965]: DEBUG nova.compute.manager [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Build of instance ab4da02e-f3ec-473b-968f-14b102780ddb was re-scheduled: Binding failed for port 71c05380-b654-4e8a-addc-e438efd4e9ca, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 828.444069] env[62965]: DEBUG nova.compute.manager [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 828.444300] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] Acquiring lock "refresh_cache-ab4da02e-f3ec-473b-968f-14b102780ddb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.444447] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] Acquired lock "refresh_cache-ab4da02e-f3ec-473b-968f-14b102780ddb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.444605] env[62965]: DEBUG nova.network.neutron [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 828.458885] env[62965]: DEBUG nova.compute.manager [req-0b4386a1-ab0f-4db9-a219-d4e6d6a5047d req-cefc19bc-5965-46e6-a480-44ee1dc08195 service nova] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Received event network-changed-e76c7578-9d13-4705-8a04-501e48bed907 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 828.459093] env[62965]: DEBUG nova.compute.manager [req-0b4386a1-ab0f-4db9-a219-d4e6d6a5047d req-cefc19bc-5965-46e6-a480-44ee1dc08195 service nova] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Refreshing instance network info cache due to event network-changed-e76c7578-9d13-4705-8a04-501e48bed907. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 828.459285] env[62965]: DEBUG oslo_concurrency.lockutils [req-0b4386a1-ab0f-4db9-a219-d4e6d6a5047d req-cefc19bc-5965-46e6-a480-44ee1dc08195 service nova] Acquiring lock "refresh_cache-3d0f74d4-ff59-462c-9577-c01c4f15ea1d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.557029] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] Releasing lock "refresh_cache-3d0f74d4-ff59-462c-9577-c01c4f15ea1d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 828.557494] env[62965]: DEBUG nova.compute.manager [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 828.557691] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 828.558011] env[62965]: DEBUG oslo_concurrency.lockutils [req-0b4386a1-ab0f-4db9-a219-d4e6d6a5047d req-cefc19bc-5965-46e6-a480-44ee1dc08195 service nova] Acquired lock "refresh_cache-3d0f74d4-ff59-462c-9577-c01c4f15ea1d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.558193] env[62965]: DEBUG nova.network.neutron [req-0b4386a1-ab0f-4db9-a219-d4e6d6a5047d req-cefc19bc-5965-46e6-a480-44ee1dc08195 service nova] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Refreshing network info cache for port e76c7578-9d13-4705-8a04-501e48bed907 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 828.559353] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f834499d-5c37-4c2b-8ba4-8aa562ab7853 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.569109] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-057e9287-2176-4a4f-aa43-81c8527ef520 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.592179] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3d0f74d4-ff59-462c-9577-c01c4f15ea1d could not be found. [ 828.592412] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 828.592656] env[62965]: INFO nova.compute.manager [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Took 0.03 seconds to destroy the instance on the hypervisor. [ 828.592822] env[62965]: DEBUG oslo.service.loopingcall [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 828.593056] env[62965]: DEBUG nova.compute.manager [-] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 828.593155] env[62965]: DEBUG nova.network.neutron [-] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 828.606372] env[62965]: DEBUG nova.network.neutron [-] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 828.777769] env[62965]: INFO nova.compute.manager [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] [instance: 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1] Took 1.02 seconds to deallocate network for instance. [ 828.962978] env[62965]: DEBUG nova.network.neutron [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 829.051715] env[62965]: DEBUG nova.network.neutron [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.076701] env[62965]: DEBUG nova.network.neutron [req-0b4386a1-ab0f-4db9-a219-d4e6d6a5047d req-cefc19bc-5965-46e6-a480-44ee1dc08195 service nova] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 829.108746] env[62965]: DEBUG nova.network.neutron [-] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.184424] env[62965]: DEBUG nova.network.neutron [req-0b4386a1-ab0f-4db9-a219-d4e6d6a5047d req-cefc19bc-5965-46e6-a480-44ee1dc08195 service nova] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.556778] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] Releasing lock "refresh_cache-ab4da02e-f3ec-473b-968f-14b102780ddb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.557043] env[62965]: DEBUG nova.compute.manager [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 829.557231] env[62965]: DEBUG nova.compute.manager [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 829.557423] env[62965]: DEBUG nova.network.neutron [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 829.574487] env[62965]: DEBUG nova.network.neutron [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 829.620288] env[62965]: INFO nova.compute.manager [-] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Took 1.03 seconds to deallocate network for instance. [ 829.623754] env[62965]: DEBUG nova.compute.claims [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 829.623991] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.683604] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae845be1-b360-4041-9d03-f49ee9b32199 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.686890] env[62965]: DEBUG oslo_concurrency.lockutils [req-0b4386a1-ab0f-4db9-a219-d4e6d6a5047d req-cefc19bc-5965-46e6-a480-44ee1dc08195 service nova] Releasing lock "refresh_cache-3d0f74d4-ff59-462c-9577-c01c4f15ea1d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.687105] env[62965]: DEBUG nova.compute.manager [req-0b4386a1-ab0f-4db9-a219-d4e6d6a5047d req-cefc19bc-5965-46e6-a480-44ee1dc08195 service nova] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Received event network-vif-deleted-e76c7578-9d13-4705-8a04-501e48bed907 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 829.692557] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e56d2299-f12b-408d-92ad-da193bd3d855 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.722673] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcf78dd6-eb3c-45f9-a0c4-a78e80e8216f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.729823] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deeb450b-5e9a-49e9-9423-5be0c9f62e30 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.742827] env[62965]: DEBUG nova.compute.provider_tree [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 829.808868] env[62965]: INFO nova.scheduler.client.report [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] Deleted allocations for instance 6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1 [ 830.078049] env[62965]: DEBUG nova.network.neutron [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.248047] env[62965]: DEBUG nova.scheduler.client.report [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 830.316897] env[62965]: DEBUG oslo_concurrency.lockutils [None req-07d5b998-5e28-4484-bef0-7af8ae66f039 tempest-AttachInterfacesV270Test-1474641000 tempest-AttachInterfacesV270Test-1474641000-project-member] Lock "6af65c7d-3bb0-47b3-a69e-14fdd1fbe1c1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 159.918s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.580946] env[62965]: INFO nova.compute.manager [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] [instance: ab4da02e-f3ec-473b-968f-14b102780ddb] Took 1.02 seconds to deallocate network for instance. [ 830.751961] env[62965]: DEBUG oslo_concurrency.lockutils [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.312s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.752515] env[62965]: DEBUG nova.compute.manager [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 830.755232] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.720s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.819561] env[62965]: DEBUG nova.compute.manager [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 831.261235] env[62965]: DEBUG nova.compute.utils [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 831.270430] env[62965]: DEBUG nova.compute.manager [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 831.272035] env[62965]: DEBUG nova.network.neutron [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 831.342535] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.363511] env[62965]: DEBUG nova.policy [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '182d38ca18c64bed8f3ecc3f95229756', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1952dedf340f4b7ba0d57af6bba9a749', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 831.555801] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04521372-ec71-460b-91bb-e4ba52230a5f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.564710] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10e1aa69-07e8-4a78-a1a0-e2b488b0d273 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.603518] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b3b3adb-4ef3-4634-b948-ad90712ee3f6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.610973] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deb8824d-31c5-4a04-acde-cc3ecb8f1e6f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.628318] env[62965]: DEBUG nova.compute.provider_tree [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 831.634923] env[62965]: INFO nova.scheduler.client.report [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] Deleted allocations for instance ab4da02e-f3ec-473b-968f-14b102780ddb [ 831.773074] env[62965]: DEBUG nova.compute.manager [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 831.953040] env[62965]: DEBUG nova.network.neutron [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Successfully created port: b7657d09-2a3b-4bf8-b674-4f37bc8ffcf4 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 832.132370] env[62965]: DEBUG nova.scheduler.client.report [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 832.144559] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e379de83-a15b-486c-9cf4-7e304061c93e tempest-ServerTagsTestJSON-407225198 tempest-ServerTagsTestJSON-407225198-project-member] Lock "ab4da02e-f3ec-473b-968f-14b102780ddb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 148.443s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.644300] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.889s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.645315] env[62965]: ERROR nova.compute.manager [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 72f714ef-8c25-45e9-b59d-636aab193fb5, please check neutron logs for more information. [ 832.645315] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Traceback (most recent call last): [ 832.645315] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 832.645315] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] self.driver.spawn(context, instance, image_meta, [ 832.645315] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 832.645315] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] self._vmops.spawn(context, instance, image_meta, injected_files, [ 832.645315] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 832.645315] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] vm_ref = self.build_virtual_machine(instance, [ 832.645315] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 832.645315] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] vif_infos = vmwarevif.get_vif_info(self._session, [ 832.645315] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 832.645638] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] for vif in network_info: [ 832.645638] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 832.645638] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] return self._sync_wrapper(fn, *args, **kwargs) [ 832.645638] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 832.645638] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] self.wait() [ 832.645638] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 832.645638] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] self[:] = self._gt.wait() [ 832.645638] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 832.645638] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] return self._exit_event.wait() [ 832.645638] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 832.645638] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] current.throw(*self._exc) [ 832.645638] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 832.645638] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] result = function(*args, **kwargs) [ 832.645985] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 832.645985] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] return func(*args, **kwargs) [ 832.645985] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 832.645985] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] raise e [ 832.645985] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 832.645985] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] nwinfo = self.network_api.allocate_for_instance( [ 832.645985] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 832.645985] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] created_port_ids = self._update_ports_for_instance( [ 832.645985] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 832.645985] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] with excutils.save_and_reraise_exception(): [ 832.645985] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 832.645985] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] self.force_reraise() [ 832.645985] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 832.646292] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] raise self.value [ 832.646292] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 832.646292] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] updated_port = self._update_port( [ 832.646292] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 832.646292] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] _ensure_no_port_binding_failure(port) [ 832.646292] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 832.646292] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] raise exception.PortBindingFailed(port_id=port['id']) [ 832.646292] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] nova.exception.PortBindingFailed: Binding failed for port 72f714ef-8c25-45e9-b59d-636aab193fb5, please check neutron logs for more information. [ 832.646292] env[62965]: ERROR nova.compute.manager [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] [ 832.646292] env[62965]: DEBUG nova.compute.utils [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Binding failed for port 72f714ef-8c25-45e9-b59d-636aab193fb5, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 832.652252] env[62965]: DEBUG nova.compute.manager [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Build of instance c8402430-4229-4b61-9118-ddeab4d8cd79 was re-scheduled: Binding failed for port 72f714ef-8c25-45e9-b59d-636aab193fb5, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 832.652836] env[62965]: DEBUG nova.compute.manager [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 832.653056] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] Acquiring lock "refresh_cache-c8402430-4229-4b61-9118-ddeab4d8cd79" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.653215] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] Acquired lock "refresh_cache-c8402430-4229-4b61-9118-ddeab4d8cd79" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.653380] env[62965]: DEBUG nova.network.neutron [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 832.657366] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.375s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.660693] env[62965]: DEBUG nova.compute.manager [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 832.784672] env[62965]: DEBUG nova.compute.manager [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 832.817281] env[62965]: DEBUG nova.virt.hardware [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 832.817964] env[62965]: DEBUG nova.virt.hardware [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 832.817964] env[62965]: DEBUG nova.virt.hardware [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 832.817964] env[62965]: DEBUG nova.virt.hardware [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 832.818206] env[62965]: DEBUG nova.virt.hardware [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 832.818388] env[62965]: DEBUG nova.virt.hardware [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 832.818608] env[62965]: DEBUG nova.virt.hardware [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 832.818825] env[62965]: DEBUG nova.virt.hardware [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 832.819057] env[62965]: DEBUG nova.virt.hardware [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 832.819386] env[62965]: DEBUG nova.virt.hardware [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 832.819530] env[62965]: DEBUG nova.virt.hardware [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 832.820581] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d023f29-1d77-49d1-9d1c-b55c856f873b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.828958] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11f76c6d-8eae-4d07-9b81-15004cf00ea7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.913823] env[62965]: DEBUG nova.compute.manager [req-278c68b6-cc8f-4f75-8db3-f2b145901fc0 req-094a9852-466f-4e48-845a-028d98e23548 service nova] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Received event network-changed-b7657d09-2a3b-4bf8-b674-4f37bc8ffcf4 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 832.914025] env[62965]: DEBUG nova.compute.manager [req-278c68b6-cc8f-4f75-8db3-f2b145901fc0 req-094a9852-466f-4e48-845a-028d98e23548 service nova] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Refreshing instance network info cache due to event network-changed-b7657d09-2a3b-4bf8-b674-4f37bc8ffcf4. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 832.914248] env[62965]: DEBUG oslo_concurrency.lockutils [req-278c68b6-cc8f-4f75-8db3-f2b145901fc0 req-094a9852-466f-4e48-845a-028d98e23548 service nova] Acquiring lock "refresh_cache-9566eb77-5198-47b3-a044-d2ed34e00888" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.914394] env[62965]: DEBUG oslo_concurrency.lockutils [req-278c68b6-cc8f-4f75-8db3-f2b145901fc0 req-094a9852-466f-4e48-845a-028d98e23548 service nova] Acquired lock "refresh_cache-9566eb77-5198-47b3-a044-d2ed34e00888" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.914557] env[62965]: DEBUG nova.network.neutron [req-278c68b6-cc8f-4f75-8db3-f2b145901fc0 req-094a9852-466f-4e48-845a-028d98e23548 service nova] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Refreshing network info cache for port b7657d09-2a3b-4bf8-b674-4f37bc8ffcf4 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 833.180672] env[62965]: ERROR nova.compute.manager [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b7657d09-2a3b-4bf8-b674-4f37bc8ffcf4, please check neutron logs for more information. [ 833.180672] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 833.180672] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 833.180672] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 833.180672] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 833.180672] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 833.180672] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 833.180672] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 833.180672] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 833.180672] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 833.180672] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 833.180672] env[62965]: ERROR nova.compute.manager raise self.value [ 833.180672] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 833.180672] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 833.180672] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 833.180672] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 833.181317] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 833.181317] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 833.181317] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b7657d09-2a3b-4bf8-b674-4f37bc8ffcf4, please check neutron logs for more information. [ 833.181317] env[62965]: ERROR nova.compute.manager [ 833.181317] env[62965]: Traceback (most recent call last): [ 833.181317] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 833.181317] env[62965]: listener.cb(fileno) [ 833.181317] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 833.181317] env[62965]: result = function(*args, **kwargs) [ 833.181317] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 833.181317] env[62965]: return func(*args, **kwargs) [ 833.181317] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 833.181317] env[62965]: raise e [ 833.181317] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 833.181317] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 833.181317] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 833.181317] env[62965]: created_port_ids = self._update_ports_for_instance( [ 833.181317] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 833.181317] env[62965]: with excutils.save_and_reraise_exception(): [ 833.181317] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 833.181317] env[62965]: self.force_reraise() [ 833.181317] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 833.181317] env[62965]: raise self.value [ 833.181317] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 833.181317] env[62965]: updated_port = self._update_port( [ 833.181317] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 833.181317] env[62965]: _ensure_no_port_binding_failure(port) [ 833.181317] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 833.181317] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 833.182199] env[62965]: nova.exception.PortBindingFailed: Binding failed for port b7657d09-2a3b-4bf8-b674-4f37bc8ffcf4, please check neutron logs for more information. [ 833.182199] env[62965]: Removing descriptor: 17 [ 833.182199] env[62965]: ERROR nova.compute.manager [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b7657d09-2a3b-4bf8-b674-4f37bc8ffcf4, please check neutron logs for more information. [ 833.182199] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Traceback (most recent call last): [ 833.182199] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 833.182199] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] yield resources [ 833.182199] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 833.182199] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] self.driver.spawn(context, instance, image_meta, [ 833.182199] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 833.182199] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] self._vmops.spawn(context, instance, image_meta, injected_files, [ 833.182199] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 833.182199] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] vm_ref = self.build_virtual_machine(instance, [ 833.182541] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 833.182541] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] vif_infos = vmwarevif.get_vif_info(self._session, [ 833.182541] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 833.182541] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] for vif in network_info: [ 833.182541] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 833.182541] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] return self._sync_wrapper(fn, *args, **kwargs) [ 833.182541] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 833.182541] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] self.wait() [ 833.182541] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 833.182541] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] self[:] = self._gt.wait() [ 833.182541] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 833.182541] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] return self._exit_event.wait() [ 833.182541] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 833.182934] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] result = hub.switch() [ 833.182934] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 833.182934] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] return self.greenlet.switch() [ 833.182934] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 833.182934] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] result = function(*args, **kwargs) [ 833.182934] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 833.182934] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] return func(*args, **kwargs) [ 833.182934] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 833.182934] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] raise e [ 833.182934] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 833.182934] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] nwinfo = self.network_api.allocate_for_instance( [ 833.182934] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 833.182934] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] created_port_ids = self._update_ports_for_instance( [ 833.184379] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 833.184379] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] with excutils.save_and_reraise_exception(): [ 833.184379] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 833.184379] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] self.force_reraise() [ 833.184379] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 833.184379] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] raise self.value [ 833.184379] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 833.184379] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] updated_port = self._update_port( [ 833.184379] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 833.184379] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] _ensure_no_port_binding_failure(port) [ 833.184379] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 833.184379] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] raise exception.PortBindingFailed(port_id=port['id']) [ 833.184689] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] nova.exception.PortBindingFailed: Binding failed for port b7657d09-2a3b-4bf8-b674-4f37bc8ffcf4, please check neutron logs for more information. [ 833.184689] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] [ 833.184689] env[62965]: INFO nova.compute.manager [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Terminating instance [ 833.184689] env[62965]: DEBUG nova.network.neutron [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 833.194553] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.310800] env[62965]: DEBUG nova.network.neutron [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.420391] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecd1a436-ccc9-400e-aacb-93c925b9c063 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.429231] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9317c257-39c7-4822-8e64-429db12cfd12 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.464022] env[62965]: DEBUG nova.network.neutron [req-278c68b6-cc8f-4f75-8db3-f2b145901fc0 req-094a9852-466f-4e48-845a-028d98e23548 service nova] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 833.464022] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fe2fa2c-7505-443d-8aa3-116300a26abf {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.472837] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aba63320-4084-40c4-b1f0-69ae0f723ae7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.487236] env[62965]: DEBUG nova.compute.provider_tree [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 833.586051] env[62965]: DEBUG nova.network.neutron [req-278c68b6-cc8f-4f75-8db3-f2b145901fc0 req-094a9852-466f-4e48-845a-028d98e23548 service nova] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.689014] env[62965]: DEBUG oslo_concurrency.lockutils [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "refresh_cache-9566eb77-5198-47b3-a044-d2ed34e00888" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.812494] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] Releasing lock "refresh_cache-c8402430-4229-4b61-9118-ddeab4d8cd79" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.812797] env[62965]: DEBUG nova.compute.manager [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 833.813029] env[62965]: DEBUG nova.compute.manager [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 833.813246] env[62965]: DEBUG nova.network.neutron [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 833.828805] env[62965]: DEBUG nova.network.neutron [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 833.991041] env[62965]: DEBUG nova.scheduler.client.report [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 834.088837] env[62965]: DEBUG oslo_concurrency.lockutils [req-278c68b6-cc8f-4f75-8db3-f2b145901fc0 req-094a9852-466f-4e48-845a-028d98e23548 service nova] Releasing lock "refresh_cache-9566eb77-5198-47b3-a044-d2ed34e00888" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.089545] env[62965]: DEBUG oslo_concurrency.lockutils [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquired lock "refresh_cache-9566eb77-5198-47b3-a044-d2ed34e00888" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.089545] env[62965]: DEBUG nova.network.neutron [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 834.338696] env[62965]: DEBUG nova.network.neutron [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.495927] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.841s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.496594] env[62965]: ERROR nova.compute.manager [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9c95488d-c11e-4946-b03a-1d15547e81a1, please check neutron logs for more information. [ 834.496594] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Traceback (most recent call last): [ 834.496594] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 834.496594] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] self.driver.spawn(context, instance, image_meta, [ 834.496594] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 834.496594] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 834.496594] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 834.496594] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] vm_ref = self.build_virtual_machine(instance, [ 834.496594] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 834.496594] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] vif_infos = vmwarevif.get_vif_info(self._session, [ 834.496594] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 834.496946] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] for vif in network_info: [ 834.496946] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 834.496946] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] return self._sync_wrapper(fn, *args, **kwargs) [ 834.496946] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 834.496946] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] self.wait() [ 834.496946] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 834.496946] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] self[:] = self._gt.wait() [ 834.496946] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 834.496946] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] return self._exit_event.wait() [ 834.496946] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 834.496946] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] current.throw(*self._exc) [ 834.496946] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 834.496946] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] result = function(*args, **kwargs) [ 834.497531] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 834.497531] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] return func(*args, **kwargs) [ 834.497531] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 834.497531] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] raise e [ 834.497531] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 834.497531] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] nwinfo = self.network_api.allocate_for_instance( [ 834.497531] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 834.497531] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] created_port_ids = self._update_ports_for_instance( [ 834.497531] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 834.497531] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] with excutils.save_and_reraise_exception(): [ 834.497531] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 834.497531] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] self.force_reraise() [ 834.497531] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 834.497954] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] raise self.value [ 834.497954] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 834.497954] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] updated_port = self._update_port( [ 834.497954] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 834.497954] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] _ensure_no_port_binding_failure(port) [ 834.497954] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 834.497954] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] raise exception.PortBindingFailed(port_id=port['id']) [ 834.497954] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] nova.exception.PortBindingFailed: Binding failed for port 9c95488d-c11e-4946-b03a-1d15547e81a1, please check neutron logs for more information. [ 834.497954] env[62965]: ERROR nova.compute.manager [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] [ 834.497954] env[62965]: DEBUG nova.compute.utils [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Binding failed for port 9c95488d-c11e-4946-b03a-1d15547e81a1, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 834.498626] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.045s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.502381] env[62965]: DEBUG nova.compute.manager [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Build of instance a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7 was re-scheduled: Binding failed for port 9c95488d-c11e-4946-b03a-1d15547e81a1, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 834.502869] env[62965]: DEBUG nova.compute.manager [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 834.503078] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] Acquiring lock "refresh_cache-a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.503227] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] Acquired lock "refresh_cache-a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.503389] env[62965]: DEBUG nova.network.neutron [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 834.610440] env[62965]: DEBUG nova.network.neutron [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 834.699676] env[62965]: DEBUG nova.network.neutron [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.842840] env[62965]: INFO nova.compute.manager [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] [instance: c8402430-4229-4b61-9118-ddeab4d8cd79] Took 1.03 seconds to deallocate network for instance. [ 834.939540] env[62965]: DEBUG nova.compute.manager [req-951a5db1-506e-4a5e-9e72-e8dd8e12de94 req-a8eb631b-7872-4059-ae93-98b0ae218e2c service nova] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Received event network-vif-deleted-b7657d09-2a3b-4bf8-b674-4f37bc8ffcf4 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 835.021447] env[62965]: DEBUG nova.network.neutron [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 835.092463] env[62965]: DEBUG nova.network.neutron [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.188414] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4dd5406-5e8c-4731-8cdd-57db0816fcf0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.196241] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab4d6c91-097e-4641-b55d-b1c9015bc9fe {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.201806] env[62965]: DEBUG oslo_concurrency.lockutils [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Releasing lock "refresh_cache-9566eb77-5198-47b3-a044-d2ed34e00888" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.202216] env[62965]: DEBUG nova.compute.manager [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 835.202404] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 835.227598] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9e41d7f2-10d4-448e-abf7-0653abb2a7ce {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.230221] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c11f12fa-6ba6-4a7b-835a-ec9d4fcf03a3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.238470] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ace65c40-8150-45a0-b452-a872fc3c7b79 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.244809] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7e03701-8c77-4372-a33d-f0acdd11f0f6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.268011] env[62965]: DEBUG nova.compute.provider_tree [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 835.273720] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9566eb77-5198-47b3-a044-d2ed34e00888 could not be found. [ 835.273960] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 835.274193] env[62965]: INFO nova.compute.manager [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Took 0.07 seconds to destroy the instance on the hypervisor. [ 835.274468] env[62965]: DEBUG oslo.service.loopingcall [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 835.274932] env[62965]: DEBUG nova.compute.manager [-] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 835.275062] env[62965]: DEBUG nova.network.neutron [-] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 835.291277] env[62965]: DEBUG nova.network.neutron [-] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 835.595459] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] Releasing lock "refresh_cache-a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.595747] env[62965]: DEBUG nova.compute.manager [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 835.595879] env[62965]: DEBUG nova.compute.manager [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 835.596059] env[62965]: DEBUG nova.network.neutron [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 835.610856] env[62965]: DEBUG nova.network.neutron [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 835.777042] env[62965]: DEBUG nova.scheduler.client.report [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 835.793221] env[62965]: DEBUG nova.network.neutron [-] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.869836] env[62965]: INFO nova.scheduler.client.report [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] Deleted allocations for instance c8402430-4229-4b61-9118-ddeab4d8cd79 [ 836.113621] env[62965]: DEBUG nova.network.neutron [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.281345] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.783s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.282033] env[62965]: ERROR nova.compute.manager [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port feea997a-7e87-460c-93a4-134225451c55, please check neutron logs for more information. [ 836.282033] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Traceback (most recent call last): [ 836.282033] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 836.282033] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] self.driver.spawn(context, instance, image_meta, [ 836.282033] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 836.282033] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 836.282033] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 836.282033] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] vm_ref = self.build_virtual_machine(instance, [ 836.282033] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 836.282033] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] vif_infos = vmwarevif.get_vif_info(self._session, [ 836.282033] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 836.282323] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] for vif in network_info: [ 836.282323] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 836.282323] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] return self._sync_wrapper(fn, *args, **kwargs) [ 836.282323] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 836.282323] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] self.wait() [ 836.282323] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 836.282323] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] self[:] = self._gt.wait() [ 836.282323] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 836.282323] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] return self._exit_event.wait() [ 836.282323] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 836.282323] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] current.throw(*self._exc) [ 836.282323] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 836.282323] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] result = function(*args, **kwargs) [ 836.282622] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 836.282622] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] return func(*args, **kwargs) [ 836.282622] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 836.282622] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] raise e [ 836.282622] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 836.282622] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] nwinfo = self.network_api.allocate_for_instance( [ 836.282622] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 836.282622] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] created_port_ids = self._update_ports_for_instance( [ 836.282622] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 836.282622] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] with excutils.save_and_reraise_exception(): [ 836.282622] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 836.282622] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] self.force_reraise() [ 836.282622] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 836.282976] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] raise self.value [ 836.282976] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 836.282976] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] updated_port = self._update_port( [ 836.282976] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 836.282976] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] _ensure_no_port_binding_failure(port) [ 836.282976] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 836.282976] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] raise exception.PortBindingFailed(port_id=port['id']) [ 836.282976] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] nova.exception.PortBindingFailed: Binding failed for port feea997a-7e87-460c-93a4-134225451c55, please check neutron logs for more information. [ 836.282976] env[62965]: ERROR nova.compute.manager [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] [ 836.282976] env[62965]: DEBUG nova.compute.utils [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Binding failed for port feea997a-7e87-460c-93a4-134225451c55, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 836.284038] env[62965]: DEBUG oslo_concurrency.lockutils [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.407s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.287184] env[62965]: DEBUG nova.compute.manager [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Build of instance 21bf78fb-4da7-4484-9678-b22a6d944c0b was re-scheduled: Binding failed for port feea997a-7e87-460c-93a4-134225451c55, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 836.287571] env[62965]: DEBUG nova.compute.manager [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 836.287794] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquiring lock "refresh_cache-21bf78fb-4da7-4484-9678-b22a6d944c0b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.287939] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquired lock "refresh_cache-21bf78fb-4da7-4484-9678-b22a6d944c0b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.288113] env[62965]: DEBUG nova.network.neutron [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 836.295266] env[62965]: INFO nova.compute.manager [-] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Took 1.02 seconds to deallocate network for instance. [ 836.297071] env[62965]: DEBUG nova.compute.claims [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 836.297269] env[62965]: DEBUG oslo_concurrency.lockutils [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.377277] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5e0aadc0-c75c-43ed-8f09-f6429e092462 tempest-ServerActionsV293TestJSON-1317416566 tempest-ServerActionsV293TestJSON-1317416566-project-member] Lock "c8402430-4229-4b61-9118-ddeab4d8cd79" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 152.165s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.618657] env[62965]: INFO nova.compute.manager [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] [instance: a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7] Took 1.02 seconds to deallocate network for instance. [ 836.810122] env[62965]: DEBUG nova.network.neutron [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 836.880055] env[62965]: DEBUG nova.compute.manager [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 836.925626] env[62965]: DEBUG nova.network.neutron [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.990029] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13037fda-beaa-418a-a3d9-6c7dbccd77e8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.997288] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29f791e7-0960-4350-825d-0d2d4bd03609 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.026944] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29e04d56-d4f5-48f3-ac08-344e56093a6a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.033982] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14a81048-aee0-4a0e-b424-0c173e7d7e6a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.047889] env[62965]: DEBUG nova.compute.provider_tree [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 837.404723] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.430204] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Releasing lock "refresh_cache-21bf78fb-4da7-4484-9678-b22a6d944c0b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.430204] env[62965]: DEBUG nova.compute.manager [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 837.430204] env[62965]: DEBUG nova.compute.manager [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 837.430204] env[62965]: DEBUG nova.network.neutron [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 837.444603] env[62965]: DEBUG nova.network.neutron [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 837.553017] env[62965]: DEBUG nova.scheduler.client.report [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 837.649371] env[62965]: INFO nova.scheduler.client.report [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] Deleted allocations for instance a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7 [ 837.947527] env[62965]: DEBUG nova.network.neutron [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.056514] env[62965]: DEBUG oslo_concurrency.lockutils [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.772s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.057165] env[62965]: ERROR nova.compute.manager [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6d23ec55-580a-4388-b3ad-9bfa3d9b47bb, please check neutron logs for more information. [ 838.057165] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Traceback (most recent call last): [ 838.057165] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 838.057165] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] self.driver.spawn(context, instance, image_meta, [ 838.057165] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 838.057165] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 838.057165] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 838.057165] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] vm_ref = self.build_virtual_machine(instance, [ 838.057165] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 838.057165] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] vif_infos = vmwarevif.get_vif_info(self._session, [ 838.057165] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 838.057467] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] for vif in network_info: [ 838.057467] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 838.057467] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] return self._sync_wrapper(fn, *args, **kwargs) [ 838.057467] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 838.057467] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] self.wait() [ 838.057467] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 838.057467] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] self[:] = self._gt.wait() [ 838.057467] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 838.057467] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] return self._exit_event.wait() [ 838.057467] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 838.057467] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] current.throw(*self._exc) [ 838.057467] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 838.057467] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] result = function(*args, **kwargs) [ 838.057760] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 838.057760] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] return func(*args, **kwargs) [ 838.057760] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 838.057760] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] raise e [ 838.057760] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 838.057760] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] nwinfo = self.network_api.allocate_for_instance( [ 838.057760] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 838.057760] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] created_port_ids = self._update_ports_for_instance( [ 838.057760] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 838.057760] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] with excutils.save_and_reraise_exception(): [ 838.057760] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 838.057760] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] self.force_reraise() [ 838.057760] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 838.058065] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] raise self.value [ 838.058065] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 838.058065] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] updated_port = self._update_port( [ 838.058065] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 838.058065] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] _ensure_no_port_binding_failure(port) [ 838.058065] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 838.058065] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] raise exception.PortBindingFailed(port_id=port['id']) [ 838.058065] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] nova.exception.PortBindingFailed: Binding failed for port 6d23ec55-580a-4388-b3ad-9bfa3d9b47bb, please check neutron logs for more information. [ 838.058065] env[62965]: ERROR nova.compute.manager [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] [ 838.058065] env[62965]: DEBUG nova.compute.utils [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Binding failed for port 6d23ec55-580a-4388-b3ad-9bfa3d9b47bb, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 838.059584] env[62965]: DEBUG nova.compute.manager [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Build of instance 13d20f82-fa4f-41ab-b08f-5b64da7dddc0 was re-scheduled: Binding failed for port 6d23ec55-580a-4388-b3ad-9bfa3d9b47bb, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 838.060086] env[62965]: DEBUG nova.compute.manager [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 838.060636] env[62965]: DEBUG oslo_concurrency.lockutils [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] Acquiring lock "refresh_cache-13d20f82-fa4f-41ab-b08f-5b64da7dddc0" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.060636] env[62965]: DEBUG oslo_concurrency.lockutils [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] Acquired lock "refresh_cache-13d20f82-fa4f-41ab-b08f-5b64da7dddc0" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.060781] env[62965]: DEBUG nova.network.neutron [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 838.061906] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.617s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.156247] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c80f522-be5f-4609-a47f-01c268c4886c tempest-ServerRescueTestJSONUnderV235-1512057181 tempest-ServerRescueTestJSONUnderV235-1512057181-project-member] Lock "a51518fb-36db-4f04-a1ed-e0ce8bfa7bb7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 136.434s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.450861] env[62965]: INFO nova.compute.manager [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 21bf78fb-4da7-4484-9678-b22a6d944c0b] Took 1.02 seconds to deallocate network for instance. [ 838.584416] env[62965]: DEBUG nova.network.neutron [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 838.659047] env[62965]: DEBUG nova.compute.manager [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 838.804899] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-707901f5-124e-4b87-ae60-5c73b700cbda {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.812835] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60443501-db03-4df4-83b9-1c7d34329d55 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.846866] env[62965]: DEBUG nova.network.neutron [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.848543] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbcb81e3-b285-4dc3-8ceb-de91b5b54fc9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.856633] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90211950-6543-42af-ba76-2e485aae65b7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.871333] env[62965]: DEBUG nova.compute.provider_tree [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 839.188734] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.352271] env[62965]: DEBUG oslo_concurrency.lockutils [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] Releasing lock "refresh_cache-13d20f82-fa4f-41ab-b08f-5b64da7dddc0" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.352533] env[62965]: DEBUG nova.compute.manager [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 839.352731] env[62965]: DEBUG nova.compute.manager [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 839.352901] env[62965]: DEBUG nova.network.neutron [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 839.374841] env[62965]: DEBUG nova.scheduler.client.report [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 839.379052] env[62965]: DEBUG nova.network.neutron [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 839.481929] env[62965]: INFO nova.scheduler.client.report [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Deleted allocations for instance 21bf78fb-4da7-4484-9678-b22a6d944c0b [ 839.882711] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.821s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.883436] env[62965]: ERROR nova.compute.manager [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c868142e-9eff-4886-a48b-aca62edbd872, please check neutron logs for more information. [ 839.883436] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Traceback (most recent call last): [ 839.883436] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 839.883436] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] self.driver.spawn(context, instance, image_meta, [ 839.883436] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 839.883436] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] self._vmops.spawn(context, instance, image_meta, injected_files, [ 839.883436] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 839.883436] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] vm_ref = self.build_virtual_machine(instance, [ 839.883436] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 839.883436] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] vif_infos = vmwarevif.get_vif_info(self._session, [ 839.883436] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 839.883752] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] for vif in network_info: [ 839.883752] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 839.883752] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] return self._sync_wrapper(fn, *args, **kwargs) [ 839.883752] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 839.883752] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] self.wait() [ 839.883752] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 839.883752] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] self[:] = self._gt.wait() [ 839.883752] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 839.883752] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] return self._exit_event.wait() [ 839.883752] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 839.883752] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] current.throw(*self._exc) [ 839.883752] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 839.883752] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] result = function(*args, **kwargs) [ 839.884149] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 839.884149] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] return func(*args, **kwargs) [ 839.884149] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 839.884149] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] raise e [ 839.884149] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 839.884149] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] nwinfo = self.network_api.allocate_for_instance( [ 839.884149] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 839.884149] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] created_port_ids = self._update_ports_for_instance( [ 839.884149] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 839.884149] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] with excutils.save_and_reraise_exception(): [ 839.884149] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 839.884149] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] self.force_reraise() [ 839.884149] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 839.884446] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] raise self.value [ 839.884446] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 839.884446] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] updated_port = self._update_port( [ 839.884446] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 839.884446] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] _ensure_no_port_binding_failure(port) [ 839.884446] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 839.884446] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] raise exception.PortBindingFailed(port_id=port['id']) [ 839.884446] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] nova.exception.PortBindingFailed: Binding failed for port c868142e-9eff-4886-a48b-aca62edbd872, please check neutron logs for more information. [ 839.884446] env[62965]: ERROR nova.compute.manager [instance: 474c8faf-cf69-4296-8fba-521e5472c071] [ 839.884446] env[62965]: DEBUG nova.compute.utils [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Binding failed for port c868142e-9eff-4886-a48b-aca62edbd872, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 839.885445] env[62965]: DEBUG nova.network.neutron [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.886750] env[62965]: DEBUG nova.compute.manager [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Build of instance 474c8faf-cf69-4296-8fba-521e5472c071 was re-scheduled: Binding failed for port c868142e-9eff-4886-a48b-aca62edbd872, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 839.887162] env[62965]: DEBUG nova.compute.manager [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 839.887482] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] Acquiring lock "refresh_cache-474c8faf-cf69-4296-8fba-521e5472c071" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.887718] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] Acquired lock "refresh_cache-474c8faf-cf69-4296-8fba-521e5472c071" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.887890] env[62965]: DEBUG nova.network.neutron [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 839.888756] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.702s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.888965] env[62965]: DEBUG nova.objects.instance [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Lazy-loading 'resources' on Instance uuid 6e0a4e4b-21a2-4185-999e-69b4c74c37d7 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 839.991494] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9f025f5d-b056-4dbb-b365-544adc3f5b5f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "21bf78fb-4da7-4484-9678-b22a6d944c0b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 132.155s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.392521] env[62965]: INFO nova.compute.manager [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] [instance: 13d20f82-fa4f-41ab-b08f-5b64da7dddc0] Took 1.04 seconds to deallocate network for instance. [ 840.424984] env[62965]: DEBUG nova.network.neutron [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 840.494149] env[62965]: DEBUG nova.compute.manager [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 840.585960] env[62965]: DEBUG nova.network.neutron [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.620662] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0dad2e2-9d5a-4640-9153-a72cecc2e365 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.631030] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7db338df-07ce-4909-bdde-78e6a7d12c66 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.666060] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36a00447-2b6f-4e33-af69-392c4ea5c1a8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.674516] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7375ca55-0aad-4f5d-9bd2-5acb4ec8befc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.689731] env[62965]: DEBUG nova.compute.provider_tree [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 841.021487] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.089858] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] Releasing lock "refresh_cache-474c8faf-cf69-4296-8fba-521e5472c071" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.091099] env[62965]: DEBUG nova.compute.manager [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 841.091284] env[62965]: DEBUG nova.compute.manager [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 841.091429] env[62965]: DEBUG nova.network.neutron [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 841.119619] env[62965]: DEBUG nova.network.neutron [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 841.192691] env[62965]: DEBUG nova.scheduler.client.report [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 841.435409] env[62965]: INFO nova.scheduler.client.report [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] Deleted allocations for instance 13d20f82-fa4f-41ab-b08f-5b64da7dddc0 [ 841.628301] env[62965]: DEBUG nova.network.neutron [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.697775] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.809s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.701226] env[62965]: DEBUG oslo_concurrency.lockutils [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.089s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.717699] env[62965]: INFO nova.scheduler.client.report [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Deleted allocations for instance 6e0a4e4b-21a2-4185-999e-69b4c74c37d7 [ 841.943231] env[62965]: DEBUG oslo_concurrency.lockutils [None req-884c4777-0c5d-4214-84b9-10de76b67aca tempest-ServerAddressesTestJSON-749032058 tempest-ServerAddressesTestJSON-749032058-project-member] Lock "13d20f82-fa4f-41ab-b08f-5b64da7dddc0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 133.254s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.131760] env[62965]: INFO nova.compute.manager [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] [instance: 474c8faf-cf69-4296-8fba-521e5472c071] Took 1.04 seconds to deallocate network for instance. [ 842.232188] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3bfad618-cc9c-44ec-88cd-2532d0b94477 tempest-ServerShowV247Test-69750264 tempest-ServerShowV247Test-69750264-project-member] Lock "6e0a4e4b-21a2-4185-999e-69b4c74c37d7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.258s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.448170] env[62965]: DEBUG nova.compute.manager [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 842.461441] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4a91c2a-9c7d-4d5a-b067-7319db173d57 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.472212] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1b022ef-8809-44a6-96cc-d36e7a200ec8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.510817] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-996716b8-840a-4e26-8b91-a0d4015ededb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.518952] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f18897c-8f17-44e3-8769-17366c09362f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.533376] env[62965]: DEBUG nova.compute.provider_tree [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 842.975469] env[62965]: DEBUG oslo_concurrency.lockutils [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.042020] env[62965]: DEBUG nova.scheduler.client.report [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 843.164872] env[62965]: INFO nova.scheduler.client.report [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] Deleted allocations for instance 474c8faf-cf69-4296-8fba-521e5472c071 [ 843.364496] env[62965]: DEBUG oslo_concurrency.lockutils [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquiring lock "6def4a38-d3b8-4724-97a1-1859b239d900" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.364496] env[62965]: DEBUG oslo_concurrency.lockutils [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "6def4a38-d3b8-4724-97a1-1859b239d900" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.547580] env[62965]: DEBUG oslo_concurrency.lockutils [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.844s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.547580] env[62965]: ERROR nova.compute.manager [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f63f2a62-43a8-4fdb-9668-0dd96a228033, please check neutron logs for more information. [ 843.547580] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Traceback (most recent call last): [ 843.547580] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 843.547580] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] self.driver.spawn(context, instance, image_meta, [ 843.547580] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 843.547580] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] self._vmops.spawn(context, instance, image_meta, injected_files, [ 843.547580] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 843.547580] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] vm_ref = self.build_virtual_machine(instance, [ 843.547951] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 843.547951] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] vif_infos = vmwarevif.get_vif_info(self._session, [ 843.547951] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 843.547951] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] for vif in network_info: [ 843.547951] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 843.547951] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] return self._sync_wrapper(fn, *args, **kwargs) [ 843.547951] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 843.547951] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] self.wait() [ 843.547951] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 843.547951] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] self[:] = self._gt.wait() [ 843.547951] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 843.547951] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] return self._exit_event.wait() [ 843.547951] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 843.548307] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] current.throw(*self._exc) [ 843.548307] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 843.548307] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] result = function(*args, **kwargs) [ 843.548307] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 843.548307] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] return func(*args, **kwargs) [ 843.548307] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 843.548307] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] raise e [ 843.548307] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 843.548307] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] nwinfo = self.network_api.allocate_for_instance( [ 843.548307] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 843.548307] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] created_port_ids = self._update_ports_for_instance( [ 843.548307] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 843.548307] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] with excutils.save_and_reraise_exception(): [ 843.548685] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 843.548685] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] self.force_reraise() [ 843.548685] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 843.548685] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] raise self.value [ 843.548685] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 843.548685] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] updated_port = self._update_port( [ 843.548685] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 843.548685] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] _ensure_no_port_binding_failure(port) [ 843.548685] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 843.548685] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] raise exception.PortBindingFailed(port_id=port['id']) [ 843.548685] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] nova.exception.PortBindingFailed: Binding failed for port f63f2a62-43a8-4fdb-9668-0dd96a228033, please check neutron logs for more information. [ 843.548685] env[62965]: ERROR nova.compute.manager [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] [ 843.549030] env[62965]: DEBUG nova.compute.utils [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Binding failed for port f63f2a62-43a8-4fdb-9668-0dd96a228033, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 843.551815] env[62965]: DEBUG nova.compute.manager [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Build of instance e9cac9cf-c165-428b-9d83-e37905f42c93 was re-scheduled: Binding failed for port f63f2a62-43a8-4fdb-9668-0dd96a228033, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 843.551815] env[62965]: DEBUG nova.compute.manager [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 843.551815] env[62965]: DEBUG oslo_concurrency.lockutils [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Acquiring lock "refresh_cache-e9cac9cf-c165-428b-9d83-e37905f42c93" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.551815] env[62965]: DEBUG oslo_concurrency.lockutils [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Acquired lock "refresh_cache-e9cac9cf-c165-428b-9d83-e37905f42c93" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.552076] env[62965]: DEBUG nova.network.neutron [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 843.553340] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.929s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.676718] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8b6e1223-88c9-4715-be5b-c4dc937dc6b0 tempest-ImagesOneServerTestJSON-19807763 tempest-ImagesOneServerTestJSON-19807763-project-member] Lock "474c8faf-cf69-4296-8fba-521e5472c071" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 127.714s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.185027] env[62965]: DEBUG nova.compute.manager [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 844.187777] env[62965]: DEBUG nova.network.neutron [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 844.257118] env[62965]: DEBUG nova.network.neutron [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.290143] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c1e37ce-fe08-48c1-a1ad-24ee3b092766 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.303615] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88a8d741-cd5c-40a6-b43a-71a8fcceb610 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.341152] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d042894-9fb4-4399-b3d0-ea6d368b7933 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.352619] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78ded2ac-76af-44f6-b491-33f93abfe694 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.367764] env[62965]: DEBUG nova.compute.provider_tree [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 844.710546] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.760906] env[62965]: DEBUG oslo_concurrency.lockutils [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Releasing lock "refresh_cache-e9cac9cf-c165-428b-9d83-e37905f42c93" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.761127] env[62965]: DEBUG nova.compute.manager [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 844.761801] env[62965]: DEBUG nova.compute.manager [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 844.762268] env[62965]: DEBUG nova.network.neutron [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 844.780975] env[62965]: DEBUG nova.network.neutron [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 844.873970] env[62965]: DEBUG nova.scheduler.client.report [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 845.074650] env[62965]: DEBUG oslo_concurrency.lockutils [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquiring lock "b59d6023-a047-4b32-bd18-1c7cc05f67d7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.074842] env[62965]: DEBUG oslo_concurrency.lockutils [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "b59d6023-a047-4b32-bd18-1c7cc05f67d7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.284546] env[62965]: DEBUG nova.network.neutron [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.379147] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.826s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.379574] env[62965]: ERROR nova.compute.manager [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e76c7578-9d13-4705-8a04-501e48bed907, please check neutron logs for more information. [ 845.379574] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Traceback (most recent call last): [ 845.379574] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 845.379574] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] self.driver.spawn(context, instance, image_meta, [ 845.379574] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 845.379574] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 845.379574] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 845.379574] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] vm_ref = self.build_virtual_machine(instance, [ 845.379574] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 845.379574] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] vif_infos = vmwarevif.get_vif_info(self._session, [ 845.379574] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 845.379972] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] for vif in network_info: [ 845.379972] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 845.379972] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] return self._sync_wrapper(fn, *args, **kwargs) [ 845.379972] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 845.379972] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] self.wait() [ 845.379972] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 845.379972] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] self[:] = self._gt.wait() [ 845.379972] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 845.379972] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] return self._exit_event.wait() [ 845.379972] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 845.379972] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] current.throw(*self._exc) [ 845.379972] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 845.379972] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] result = function(*args, **kwargs) [ 845.380458] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 845.380458] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] return func(*args, **kwargs) [ 845.380458] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 845.380458] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] raise e [ 845.380458] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 845.380458] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] nwinfo = self.network_api.allocate_for_instance( [ 845.380458] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 845.380458] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] created_port_ids = self._update_ports_for_instance( [ 845.380458] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 845.380458] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] with excutils.save_and_reraise_exception(): [ 845.380458] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 845.380458] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] self.force_reraise() [ 845.380458] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 845.380769] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] raise self.value [ 845.380769] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 845.380769] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] updated_port = self._update_port( [ 845.380769] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 845.380769] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] _ensure_no_port_binding_failure(port) [ 845.380769] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 845.380769] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] raise exception.PortBindingFailed(port_id=port['id']) [ 845.380769] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] nova.exception.PortBindingFailed: Binding failed for port e76c7578-9d13-4705-8a04-501e48bed907, please check neutron logs for more information. [ 845.380769] env[62965]: ERROR nova.compute.manager [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] [ 845.380769] env[62965]: DEBUG nova.compute.utils [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Binding failed for port e76c7578-9d13-4705-8a04-501e48bed907, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 845.381834] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.040s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.383669] env[62965]: INFO nova.compute.claims [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 845.386655] env[62965]: DEBUG nova.compute.manager [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Build of instance 3d0f74d4-ff59-462c-9577-c01c4f15ea1d was re-scheduled: Binding failed for port e76c7578-9d13-4705-8a04-501e48bed907, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 845.387151] env[62965]: DEBUG nova.compute.manager [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 845.387637] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] Acquiring lock "refresh_cache-3d0f74d4-ff59-462c-9577-c01c4f15ea1d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.387637] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] Acquired lock "refresh_cache-3d0f74d4-ff59-462c-9577-c01c4f15ea1d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.388252] env[62965]: DEBUG nova.network.neutron [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 845.788108] env[62965]: INFO nova.compute.manager [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] [instance: e9cac9cf-c165-428b-9d83-e37905f42c93] Took 1.03 seconds to deallocate network for instance. [ 845.910116] env[62965]: DEBUG nova.network.neutron [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 845.999158] env[62965]: DEBUG nova.network.neutron [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.074426] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "bd36601b-5a70-4a29-8ef1-d1e925f41de7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.074689] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "bd36601b-5a70-4a29-8ef1-d1e925f41de7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.502583] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] Releasing lock "refresh_cache-3d0f74d4-ff59-462c-9577-c01c4f15ea1d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.502852] env[62965]: DEBUG nova.compute.manager [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 846.502982] env[62965]: DEBUG nova.compute.manager [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 846.503192] env[62965]: DEBUG nova.network.neutron [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 846.528923] env[62965]: DEBUG nova.network.neutron [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 846.633174] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc6896b6-097a-4aba-a843-06b09d075091 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.641413] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4e57c78-dfbd-44cb-bda0-32ddd939818d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.678903] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a270544a-f2a1-4a75-ae71-aaf345b167cc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.686546] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fbed809-e3a3-4fab-958a-34c0eb9d6840 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.700344] env[62965]: DEBUG nova.compute.provider_tree [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 846.819710] env[62965]: INFO nova.scheduler.client.report [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Deleted allocations for instance e9cac9cf-c165-428b-9d83-e37905f42c93 [ 847.031738] env[62965]: DEBUG nova.network.neutron [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.205153] env[62965]: DEBUG nova.scheduler.client.report [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 847.330866] env[62965]: DEBUG oslo_concurrency.lockutils [None req-58ce1b57-99be-4e2d-9db1-f54443f5fa80 tempest-MigrationsAdminTest-2147482180 tempest-MigrationsAdminTest-2147482180-project-member] Lock "e9cac9cf-c165-428b-9d83-e37905f42c93" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 113.679s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.536602] env[62965]: INFO nova.compute.manager [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] [instance: 3d0f74d4-ff59-462c-9577-c01c4f15ea1d] Took 1.03 seconds to deallocate network for instance. [ 847.711460] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.329s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.713199] env[62965]: DEBUG nova.compute.manager [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 847.714661] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.520s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.718237] env[62965]: INFO nova.compute.claims [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 847.836020] env[62965]: DEBUG nova.compute.manager [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 848.222281] env[62965]: DEBUG nova.compute.utils [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 848.228859] env[62965]: DEBUG nova.compute.manager [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 848.229049] env[62965]: DEBUG nova.network.neutron [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 848.297593] env[62965]: DEBUG nova.policy [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7811e3bed74f4d71857f66cb1961ad74', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c452f4bded584fdeab6d8cfd7061e63b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 848.373342] env[62965]: DEBUG oslo_concurrency.lockutils [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.580341] env[62965]: INFO nova.scheduler.client.report [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] Deleted allocations for instance 3d0f74d4-ff59-462c-9577-c01c4f15ea1d [ 848.732833] env[62965]: DEBUG nova.compute.manager [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 848.750955] env[62965]: DEBUG nova.network.neutron [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Successfully created port: 2c3621b8-53b9-48d0-90ca-6f9450d91d7a {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 848.943509] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-164de308-68fa-4ba8-a5d9-b858cdb28471 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.951915] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1038c7d8-bfe4-4226-9e34-8cf54d9f9402 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.985151] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feadbb10-c76c-496e-ba9e-0293b162021c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.993154] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2439944-9ae6-4022-b094-f68fee763a10 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.006759] env[62965]: DEBUG nova.compute.provider_tree [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 849.091324] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f232a0ff-5374-4132-ad9d-599f5af28165 tempest-ServersTestManualDisk-1309091836 tempest-ServersTestManualDisk-1309091836-project-member] Lock "3d0f74d4-ff59-462c-9577-c01c4f15ea1d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.979s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.510154] env[62965]: DEBUG nova.scheduler.client.report [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 849.593948] env[62965]: DEBUG nova.compute.manager [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 849.749020] env[62965]: DEBUG nova.compute.manager [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 849.791788] env[62965]: DEBUG nova.virt.hardware [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 849.791788] env[62965]: DEBUG nova.virt.hardware [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 849.791788] env[62965]: DEBUG nova.virt.hardware [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 849.792015] env[62965]: DEBUG nova.virt.hardware [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 849.793793] env[62965]: DEBUG nova.virt.hardware [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 849.795320] env[62965]: DEBUG nova.virt.hardware [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 849.795320] env[62965]: DEBUG nova.virt.hardware [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 849.795320] env[62965]: DEBUG nova.virt.hardware [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 849.795320] env[62965]: DEBUG nova.virt.hardware [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 849.795320] env[62965]: DEBUG nova.virt.hardware [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 849.795576] env[62965]: DEBUG nova.virt.hardware [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 849.796659] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41c99c17-8e83-4d3a-a130-dbcd82b523af {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.805664] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8482c55f-f909-4766-92bb-f2b046af10da {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.921906] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquiring lock "a0e9cc1a-bc95-4a44-99c5-aabf85a373f4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.923298] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lock "a0e9cc1a-bc95-4a44-99c5-aabf85a373f4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.002107] env[62965]: DEBUG nova.compute.manager [req-dbd29049-44e7-4216-8436-4043c8d2b8aa req-846cfefb-e294-40b2-845f-ee83169d7606 service nova] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Received event network-changed-2c3621b8-53b9-48d0-90ca-6f9450d91d7a {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 850.002186] env[62965]: DEBUG nova.compute.manager [req-dbd29049-44e7-4216-8436-4043c8d2b8aa req-846cfefb-e294-40b2-845f-ee83169d7606 service nova] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Refreshing instance network info cache due to event network-changed-2c3621b8-53b9-48d0-90ca-6f9450d91d7a. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 850.002426] env[62965]: DEBUG oslo_concurrency.lockutils [req-dbd29049-44e7-4216-8436-4043c8d2b8aa req-846cfefb-e294-40b2-845f-ee83169d7606 service nova] Acquiring lock "refresh_cache-2b6efa32-fc5b-402c-a551-b6e5c01eca2c" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.002522] env[62965]: DEBUG oslo_concurrency.lockutils [req-dbd29049-44e7-4216-8436-4043c8d2b8aa req-846cfefb-e294-40b2-845f-ee83169d7606 service nova] Acquired lock "refresh_cache-2b6efa32-fc5b-402c-a551-b6e5c01eca2c" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.002708] env[62965]: DEBUG nova.network.neutron [req-dbd29049-44e7-4216-8436-4043c8d2b8aa req-846cfefb-e294-40b2-845f-ee83169d7606 service nova] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Refreshing network info cache for port 2c3621b8-53b9-48d0-90ca-6f9450d91d7a {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 850.020102] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.302s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.020197] env[62965]: DEBUG nova.compute.manager [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 850.027684] env[62965]: DEBUG oslo_concurrency.lockutils [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.730s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.133168] env[62965]: DEBUG oslo_concurrency.lockutils [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.147334] env[62965]: ERROR nova.compute.manager [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2c3621b8-53b9-48d0-90ca-6f9450d91d7a, please check neutron logs for more information. [ 850.147334] env[62965]: ERROR nova.compute.manager Traceback (most recent call last): [ 850.147334] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 850.147334] env[62965]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 850.147334] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 850.147334] env[62965]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 850.147334] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 850.147334] env[62965]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 850.147334] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 850.147334] env[62965]: ERROR nova.compute.manager self.force_reraise() [ 850.147334] env[62965]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 850.147334] env[62965]: ERROR nova.compute.manager raise self.value [ 850.147334] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 850.147334] env[62965]: ERROR nova.compute.manager updated_port = self._update_port( [ 850.147334] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 850.147334] env[62965]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 850.147884] env[62965]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 850.147884] env[62965]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 850.147884] env[62965]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2c3621b8-53b9-48d0-90ca-6f9450d91d7a, please check neutron logs for more information. [ 850.147884] env[62965]: ERROR nova.compute.manager [ 850.147884] env[62965]: Traceback (most recent call last): [ 850.147884] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 850.147884] env[62965]: listener.cb(fileno) [ 850.147884] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 850.147884] env[62965]: result = function(*args, **kwargs) [ 850.147884] env[62965]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 850.147884] env[62965]: return func(*args, **kwargs) [ 850.147884] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 850.147884] env[62965]: raise e [ 850.147884] env[62965]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 850.147884] env[62965]: nwinfo = self.network_api.allocate_for_instance( [ 850.147884] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 850.147884] env[62965]: created_port_ids = self._update_ports_for_instance( [ 850.147884] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 850.147884] env[62965]: with excutils.save_and_reraise_exception(): [ 850.147884] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 850.147884] env[62965]: self.force_reraise() [ 850.147884] env[62965]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 850.147884] env[62965]: raise self.value [ 850.147884] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 850.147884] env[62965]: updated_port = self._update_port( [ 850.147884] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 850.147884] env[62965]: _ensure_no_port_binding_failure(port) [ 850.147884] env[62965]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 850.147884] env[62965]: raise exception.PortBindingFailed(port_id=port['id']) [ 850.148709] env[62965]: nova.exception.PortBindingFailed: Binding failed for port 2c3621b8-53b9-48d0-90ca-6f9450d91d7a, please check neutron logs for more information. [ 850.148709] env[62965]: Removing descriptor: 19 [ 850.148709] env[62965]: ERROR nova.compute.manager [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2c3621b8-53b9-48d0-90ca-6f9450d91d7a, please check neutron logs for more information. [ 850.148709] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Traceback (most recent call last): [ 850.148709] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 850.148709] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] yield resources [ 850.148709] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 850.148709] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] self.driver.spawn(context, instance, image_meta, [ 850.148709] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 850.148709] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 850.148709] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 850.148709] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] vm_ref = self.build_virtual_machine(instance, [ 850.149059] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 850.149059] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] vif_infos = vmwarevif.get_vif_info(self._session, [ 850.149059] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 850.149059] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] for vif in network_info: [ 850.149059] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 850.149059] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] return self._sync_wrapper(fn, *args, **kwargs) [ 850.149059] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 850.149059] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] self.wait() [ 850.149059] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 850.149059] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] self[:] = self._gt.wait() [ 850.149059] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 850.149059] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] return self._exit_event.wait() [ 850.149059] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 850.149409] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] result = hub.switch() [ 850.149409] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 850.149409] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] return self.greenlet.switch() [ 850.149409] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 850.149409] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] result = function(*args, **kwargs) [ 850.149409] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 850.149409] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] return func(*args, **kwargs) [ 850.149409] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 850.149409] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] raise e [ 850.149409] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 850.149409] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] nwinfo = self.network_api.allocate_for_instance( [ 850.149409] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 850.149409] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] created_port_ids = self._update_ports_for_instance( [ 850.149762] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 850.149762] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] with excutils.save_and_reraise_exception(): [ 850.149762] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 850.149762] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] self.force_reraise() [ 850.149762] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 850.149762] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] raise self.value [ 850.149762] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 850.149762] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] updated_port = self._update_port( [ 850.149762] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 850.149762] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] _ensure_no_port_binding_failure(port) [ 850.149762] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 850.149762] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] raise exception.PortBindingFailed(port_id=port['id']) [ 850.150099] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] nova.exception.PortBindingFailed: Binding failed for port 2c3621b8-53b9-48d0-90ca-6f9450d91d7a, please check neutron logs for more information. [ 850.150099] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] [ 850.150099] env[62965]: INFO nova.compute.manager [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Terminating instance [ 850.525051] env[62965]: DEBUG nova.network.neutron [req-dbd29049-44e7-4216-8436-4043c8d2b8aa req-846cfefb-e294-40b2-845f-ee83169d7606 service nova] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 850.537268] env[62965]: DEBUG nova.compute.utils [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 850.539680] env[62965]: DEBUG nova.compute.manager [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 850.539680] env[62965]: DEBUG nova.network.neutron [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 850.622151] env[62965]: DEBUG nova.network.neutron [req-dbd29049-44e7-4216-8436-4043c8d2b8aa req-846cfefb-e294-40b2-845f-ee83169d7606 service nova] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.654693] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] Acquiring lock "refresh_cache-2b6efa32-fc5b-402c-a551-b6e5c01eca2c" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.674998] env[62965]: DEBUG nova.policy [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7e9d1908a194480e9ee9f5fa32b15cc6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd9897f889b854681909285c643d1ae86', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 850.751282] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77e2e62d-4677-4951-8b57-626681e56d94 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.759139] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cc75f1c-76f7-4fdb-a4f4-4328f302c137 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.797038] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0aaea0d-e692-43b2-9490-a3010e1aa937 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.809163] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-154cffbb-f3e9-42b0-b72d-9f47c4161e62 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.824222] env[62965]: DEBUG nova.compute.provider_tree [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 851.043196] env[62965]: DEBUG nova.compute.manager [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 851.125042] env[62965]: DEBUG oslo_concurrency.lockutils [req-dbd29049-44e7-4216-8436-4043c8d2b8aa req-846cfefb-e294-40b2-845f-ee83169d7606 service nova] Releasing lock "refresh_cache-2b6efa32-fc5b-402c-a551-b6e5c01eca2c" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.125448] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] Acquired lock "refresh_cache-2b6efa32-fc5b-402c-a551-b6e5c01eca2c" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.125632] env[62965]: DEBUG nova.network.neutron [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 851.331411] env[62965]: DEBUG nova.scheduler.client.report [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 851.379554] env[62965]: DEBUG nova.network.neutron [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Successfully created port: 97930362-17f2-40a5-bc09-a9df5f010935 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 851.658521] env[62965]: DEBUG nova.network.neutron [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 851.791277] env[62965]: DEBUG nova.network.neutron [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.835431] env[62965]: DEBUG oslo_concurrency.lockutils [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.808s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.836380] env[62965]: ERROR nova.compute.manager [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b7657d09-2a3b-4bf8-b674-4f37bc8ffcf4, please check neutron logs for more information. [ 851.836380] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Traceback (most recent call last): [ 851.836380] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 851.836380] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] self.driver.spawn(context, instance, image_meta, [ 851.836380] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 851.836380] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] self._vmops.spawn(context, instance, image_meta, injected_files, [ 851.836380] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 851.836380] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] vm_ref = self.build_virtual_machine(instance, [ 851.836380] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 851.836380] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] vif_infos = vmwarevif.get_vif_info(self._session, [ 851.836380] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 851.836883] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] for vif in network_info: [ 851.836883] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 851.836883] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] return self._sync_wrapper(fn, *args, **kwargs) [ 851.836883] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 851.836883] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] self.wait() [ 851.836883] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 851.836883] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] self[:] = self._gt.wait() [ 851.836883] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 851.836883] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] return self._exit_event.wait() [ 851.836883] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 851.836883] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] result = hub.switch() [ 851.836883] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 851.836883] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] return self.greenlet.switch() [ 851.837244] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 851.837244] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] result = function(*args, **kwargs) [ 851.837244] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 851.837244] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] return func(*args, **kwargs) [ 851.837244] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 851.837244] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] raise e [ 851.837244] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 851.837244] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] nwinfo = self.network_api.allocate_for_instance( [ 851.837244] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 851.837244] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] created_port_ids = self._update_ports_for_instance( [ 851.837244] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 851.837244] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] with excutils.save_and_reraise_exception(): [ 851.837244] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 851.837669] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] self.force_reraise() [ 851.837669] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 851.837669] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] raise self.value [ 851.837669] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 851.837669] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] updated_port = self._update_port( [ 851.837669] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 851.837669] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] _ensure_no_port_binding_failure(port) [ 851.837669] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 851.837669] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] raise exception.PortBindingFailed(port_id=port['id']) [ 851.837669] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] nova.exception.PortBindingFailed: Binding failed for port b7657d09-2a3b-4bf8-b674-4f37bc8ffcf4, please check neutron logs for more information. [ 851.837669] env[62965]: ERROR nova.compute.manager [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] [ 851.837919] env[62965]: DEBUG nova.compute.utils [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Binding failed for port b7657d09-2a3b-4bf8-b674-4f37bc8ffcf4, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 851.840036] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.435s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.841171] env[62965]: INFO nova.compute.claims [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 851.845134] env[62965]: DEBUG nova.compute.manager [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Build of instance 9566eb77-5198-47b3-a044-d2ed34e00888 was re-scheduled: Binding failed for port b7657d09-2a3b-4bf8-b674-4f37bc8ffcf4, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 851.845586] env[62965]: DEBUG nova.compute.manager [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 851.845808] env[62965]: DEBUG oslo_concurrency.lockutils [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "refresh_cache-9566eb77-5198-47b3-a044-d2ed34e00888" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.845997] env[62965]: DEBUG oslo_concurrency.lockutils [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquired lock "refresh_cache-9566eb77-5198-47b3-a044-d2ed34e00888" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.846175] env[62965]: DEBUG nova.network.neutron [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 852.059580] env[62965]: DEBUG nova.compute.manager [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 852.093510] env[62965]: DEBUG nova.virt.hardware [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 852.093815] env[62965]: DEBUG nova.virt.hardware [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 852.093941] env[62965]: DEBUG nova.virt.hardware [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 852.094225] env[62965]: DEBUG nova.virt.hardware [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 852.094334] env[62965]: DEBUG nova.virt.hardware [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 852.094904] env[62965]: DEBUG nova.virt.hardware [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 852.095021] env[62965]: DEBUG nova.virt.hardware [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 852.095153] env[62965]: DEBUG nova.virt.hardware [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 852.095513] env[62965]: DEBUG nova.virt.hardware [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 852.095713] env[62965]: DEBUG nova.virt.hardware [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 852.095895] env[62965]: DEBUG nova.virt.hardware [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 852.097026] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fc2fd77-b637-4a8b-8345-7cef5c2dc396 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.104297] env[62965]: DEBUG nova.compute.manager [req-b292a8c2-079e-4637-8aad-208cf812816f req-eeca3db3-5d73-4736-8105-db06176f7e9a service nova] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Received event network-vif-deleted-2c3621b8-53b9-48d0-90ca-6f9450d91d7a {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 852.108961] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c44606f-4e62-469c-89fc-ddab5222ab17 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.300822] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] Releasing lock "refresh_cache-2b6efa32-fc5b-402c-a551-b6e5c01eca2c" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.301234] env[62965]: DEBUG nova.compute.manager [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 852.301621] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 852.301908] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d7232f61-b806-4112-ae44-4bc0ba8c3ee3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.319543] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0af5d0e7-7030-4fe3-889a-ed8c9975f54c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.353026] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2b6efa32-fc5b-402c-a551-b6e5c01eca2c could not be found. [ 852.353260] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 852.353436] env[62965]: INFO nova.compute.manager [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Took 0.05 seconds to destroy the instance on the hypervisor. [ 852.353675] env[62965]: DEBUG oslo.service.loopingcall [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 852.354144] env[62965]: DEBUG nova.compute.manager [-] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 852.354240] env[62965]: DEBUG nova.network.neutron [-] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 852.384468] env[62965]: DEBUG nova.network.neutron [-] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 852.393925] env[62965]: DEBUG nova.network.neutron [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 852.488508] env[62965]: DEBUG nova.network.neutron [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.887948] env[62965]: DEBUG nova.network.neutron [-] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.994643] env[62965]: DEBUG oslo_concurrency.lockutils [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Releasing lock "refresh_cache-9566eb77-5198-47b3-a044-d2ed34e00888" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.994831] env[62965]: DEBUG nova.compute.manager [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 852.995063] env[62965]: DEBUG nova.compute.manager [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 852.995232] env[62965]: DEBUG nova.network.neutron [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 853.016040] env[62965]: DEBUG nova.network.neutron [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 853.055465] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-051bd0a7-bbe6-439a-b794-7621a5d23f06 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.064333] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e511393a-785a-43d2-8af2-117ee05abe8c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.096633] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f02f7b8-8711-4b03-ae3c-6dd3809d2054 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.103783] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78f70a37-d5af-434d-85f5-c05fdb1d9f81 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.119025] env[62965]: DEBUG nova.compute.provider_tree [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 853.374738] env[62965]: DEBUG nova.compute.manager [req-16cb714f-ceb3-4d20-92c3-70a2d988c8c9 req-12ccbc83-06d6-4926-ad25-c04cc466828a service nova] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Received event network-vif-plugged-97930362-17f2-40a5-bc09-a9df5f010935 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 853.374738] env[62965]: DEBUG oslo_concurrency.lockutils [req-16cb714f-ceb3-4d20-92c3-70a2d988c8c9 req-12ccbc83-06d6-4926-ad25-c04cc466828a service nova] Acquiring lock "64cf898c-2c97-4daa-80c8-0a1c252cbcf2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.374738] env[62965]: DEBUG oslo_concurrency.lockutils [req-16cb714f-ceb3-4d20-92c3-70a2d988c8c9 req-12ccbc83-06d6-4926-ad25-c04cc466828a service nova] Lock "64cf898c-2c97-4daa-80c8-0a1c252cbcf2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.374738] env[62965]: DEBUG oslo_concurrency.lockutils [req-16cb714f-ceb3-4d20-92c3-70a2d988c8c9 req-12ccbc83-06d6-4926-ad25-c04cc466828a service nova] Lock "64cf898c-2c97-4daa-80c8-0a1c252cbcf2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.374738] env[62965]: DEBUG nova.compute.manager [req-16cb714f-ceb3-4d20-92c3-70a2d988c8c9 req-12ccbc83-06d6-4926-ad25-c04cc466828a service nova] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] No waiting events found dispatching network-vif-plugged-97930362-17f2-40a5-bc09-a9df5f010935 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 853.374951] env[62965]: WARNING nova.compute.manager [req-16cb714f-ceb3-4d20-92c3-70a2d988c8c9 req-12ccbc83-06d6-4926-ad25-c04cc466828a service nova] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Received unexpected event network-vif-plugged-97930362-17f2-40a5-bc09-a9df5f010935 for instance with vm_state building and task_state spawning. [ 853.394022] env[62965]: INFO nova.compute.manager [-] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Took 1.04 seconds to deallocate network for instance. [ 853.394022] env[62965]: DEBUG nova.compute.claims [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Aborting claim: {{(pid=62965) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 853.394556] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.431531] env[62965]: DEBUG nova.network.neutron [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Successfully updated port: 97930362-17f2-40a5-bc09-a9df5f010935 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 853.521078] env[62965]: DEBUG nova.network.neutron [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.622502] env[62965]: DEBUG nova.scheduler.client.report [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 853.938828] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Acquiring lock "refresh_cache-64cf898c-2c97-4daa-80c8-0a1c252cbcf2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.938828] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Acquired lock "refresh_cache-64cf898c-2c97-4daa-80c8-0a1c252cbcf2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.939302] env[62965]: DEBUG nova.network.neutron [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 854.026105] env[62965]: INFO nova.compute.manager [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 9566eb77-5198-47b3-a044-d2ed34e00888] Took 1.03 seconds to deallocate network for instance. [ 854.127187] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.287s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.127736] env[62965]: DEBUG nova.compute.manager [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 854.130538] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.942s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.131998] env[62965]: INFO nova.compute.claims [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 854.483867] env[62965]: DEBUG nova.network.neutron [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 854.637642] env[62965]: DEBUG nova.compute.utils [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 854.641149] env[62965]: DEBUG nova.compute.manager [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Not allocating networking since 'none' was specified. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 854.827942] env[62965]: DEBUG nova.network.neutron [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Updating instance_info_cache with network_info: [{"id": "97930362-17f2-40a5-bc09-a9df5f010935", "address": "fa:16:3e:a4:ce:72", "network": {"id": "88959f71-7c4c-4c1d-aef7-b674c749fff6", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-692265432-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d9897f889b854681909285c643d1ae86", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2d2742ba-c3af-4412-877d-c2811dfeba46", "external-id": "nsx-vlan-transportzone-390", "segmentation_id": 390, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97930362-17", "ovs_interfaceid": "97930362-17f2-40a5-bc09-a9df5f010935", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.066453] env[62965]: INFO nova.scheduler.client.report [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Deleted allocations for instance 9566eb77-5198-47b3-a044-d2ed34e00888 [ 855.142417] env[62965]: DEBUG nova.compute.manager [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 855.330464] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Releasing lock "refresh_cache-64cf898c-2c97-4daa-80c8-0a1c252cbcf2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.331852] env[62965]: DEBUG nova.compute.manager [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Instance network_info: |[{"id": "97930362-17f2-40a5-bc09-a9df5f010935", "address": "fa:16:3e:a4:ce:72", "network": {"id": "88959f71-7c4c-4c1d-aef7-b674c749fff6", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-692265432-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d9897f889b854681909285c643d1ae86", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2d2742ba-c3af-4412-877d-c2811dfeba46", "external-id": "nsx-vlan-transportzone-390", "segmentation_id": 390, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97930362-17", "ovs_interfaceid": "97930362-17f2-40a5-bc09-a9df5f010935", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 855.332135] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a4:ce:72', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2d2742ba-c3af-4412-877d-c2811dfeba46', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '97930362-17f2-40a5-bc09-a9df5f010935', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 855.339882] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Creating folder: Project (d9897f889b854681909285c643d1ae86). Parent ref: group-v294931. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 855.342673] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2d89a0e6-e41d-40b1-bd2f-fe8ecf2f4bcd {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.353499] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Created folder: Project (d9897f889b854681909285c643d1ae86) in parent group-v294931. [ 855.353676] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Creating folder: Instances. Parent ref: group-v294953. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 855.357020] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f6cff63e-7448-44ed-ba97-d279e97a730d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.359667] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85e3d833-f53f-48bf-bc8e-1f9c660134fa {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.363433] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Created folder: Instances in parent group-v294953. [ 855.363657] env[62965]: DEBUG oslo.service.loopingcall [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 855.364203] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 855.364424] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8defd534-93a6-454d-80c2-a3cd01f4dd2f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.384841] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e53968e-96da-45c9-a11b-e46b95665ef7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.388784] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 855.388784] env[62965]: value = "task-1389817" [ 855.388784] env[62965]: _type = "Task" [ 855.388784] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.420183] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41cc70d8-bbaa-466c-9eee-068523316e90 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.426149] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389817, 'name': CreateVM_Task} progress is 15%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.433410] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b308804f-f6d5-414f-ab80-01f59527eaa5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.446888] env[62965]: DEBUG nova.compute.provider_tree [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 855.451875] env[62965]: DEBUG nova.compute.manager [req-672ad77f-7f36-4617-8670-48c5f56e8b13 req-035a6ee2-b8a1-423d-a9ea-f267f2a8bf7a service nova] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Received event network-changed-97930362-17f2-40a5-bc09-a9df5f010935 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 855.452038] env[62965]: DEBUG nova.compute.manager [req-672ad77f-7f36-4617-8670-48c5f56e8b13 req-035a6ee2-b8a1-423d-a9ea-f267f2a8bf7a service nova] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Refreshing instance network info cache due to event network-changed-97930362-17f2-40a5-bc09-a9df5f010935. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 855.452264] env[62965]: DEBUG oslo_concurrency.lockutils [req-672ad77f-7f36-4617-8670-48c5f56e8b13 req-035a6ee2-b8a1-423d-a9ea-f267f2a8bf7a service nova] Acquiring lock "refresh_cache-64cf898c-2c97-4daa-80c8-0a1c252cbcf2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.452696] env[62965]: DEBUG oslo_concurrency.lockutils [req-672ad77f-7f36-4617-8670-48c5f56e8b13 req-035a6ee2-b8a1-423d-a9ea-f267f2a8bf7a service nova] Acquired lock "refresh_cache-64cf898c-2c97-4daa-80c8-0a1c252cbcf2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.452696] env[62965]: DEBUG nova.network.neutron [req-672ad77f-7f36-4617-8670-48c5f56e8b13 req-035a6ee2-b8a1-423d-a9ea-f267f2a8bf7a service nova] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Refreshing network info cache for port 97930362-17f2-40a5-bc09-a9df5f010935 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 855.577469] env[62965]: DEBUG oslo_concurrency.lockutils [None req-48cdf510-94c2-4e74-a0fc-831c0c717751 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "9566eb77-5198-47b3-a044-d2ed34e00888" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.929s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.899570] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389817, 'name': CreateVM_Task, 'duration_secs': 0.334985} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.902014] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 855.911612] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.911612] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.911728] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 855.912073] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e60588dd-186e-42f9-917d-18d077fd820b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.916706] env[62965]: DEBUG oslo_vmware.api [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Waiting for the task: (returnval){ [ 855.916706] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5223f2a6-6244-60e2-d090-ba0f53c5cd10" [ 855.916706] env[62965]: _type = "Task" [ 855.916706] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.924271] env[62965]: DEBUG oslo_vmware.api [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5223f2a6-6244-60e2-d090-ba0f53c5cd10, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.950464] env[62965]: DEBUG nova.scheduler.client.report [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 856.081027] env[62965]: DEBUG nova.compute.manager [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 856.157635] env[62965]: DEBUG nova.compute.manager [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 856.192155] env[62965]: DEBUG nova.virt.hardware [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 856.192388] env[62965]: DEBUG nova.virt.hardware [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 856.192537] env[62965]: DEBUG nova.virt.hardware [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 856.192707] env[62965]: DEBUG nova.virt.hardware [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 856.192848] env[62965]: DEBUG nova.virt.hardware [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 856.192986] env[62965]: DEBUG nova.virt.hardware [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 856.193206] env[62965]: DEBUG nova.virt.hardware [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 856.193425] env[62965]: DEBUG nova.virt.hardware [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 856.193527] env[62965]: DEBUG nova.virt.hardware [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 856.193690] env[62965]: DEBUG nova.virt.hardware [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 856.193859] env[62965]: DEBUG nova.virt.hardware [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 856.194756] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9d26f0b-e7b7-4c32-b0c5-c826560a3ebf {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.205948] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1781df6-a444-485b-98de-22d4463359c5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.222918] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Instance VIF info [] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 856.232660] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Creating folder: Project (692a6aeac7904a0cbd6c49c0edb26395). Parent ref: group-v294931. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 856.232660] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-86910b99-a4de-403e-af3c-4ea08ac507ce {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.242317] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Created folder: Project (692a6aeac7904a0cbd6c49c0edb26395) in parent group-v294931. [ 856.242317] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Creating folder: Instances. Parent ref: group-v294956. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 856.242391] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fcdac71c-034c-47dc-bf8c-1bc553828b4a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.251291] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Created folder: Instances in parent group-v294956. [ 856.251535] env[62965]: DEBUG oslo.service.loopingcall [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 856.251871] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 856.251996] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-787c3865-24d1-4b0b-a4ec-5283a52059cd {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.269969] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 856.269969] env[62965]: value = "task-1389820" [ 856.269969] env[62965]: _type = "Task" [ 856.269969] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.281565] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389820, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.358410] env[62965]: DEBUG nova.network.neutron [req-672ad77f-7f36-4617-8670-48c5f56e8b13 req-035a6ee2-b8a1-423d-a9ea-f267f2a8bf7a service nova] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Updated VIF entry in instance network info cache for port 97930362-17f2-40a5-bc09-a9df5f010935. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 856.358697] env[62965]: DEBUG nova.network.neutron [req-672ad77f-7f36-4617-8670-48c5f56e8b13 req-035a6ee2-b8a1-423d-a9ea-f267f2a8bf7a service nova] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Updating instance_info_cache with network_info: [{"id": "97930362-17f2-40a5-bc09-a9df5f010935", "address": "fa:16:3e:a4:ce:72", "network": {"id": "88959f71-7c4c-4c1d-aef7-b674c749fff6", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-692265432-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d9897f889b854681909285c643d1ae86", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2d2742ba-c3af-4412-877d-c2811dfeba46", "external-id": "nsx-vlan-transportzone-390", "segmentation_id": 390, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97930362-17", "ovs_interfaceid": "97930362-17f2-40a5-bc09-a9df5f010935", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.417847] env[62965]: DEBUG oslo_concurrency.lockutils [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "4fc543a1-34ce-40bb-9e80-72ab9098eb38" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.418130] env[62965]: DEBUG oslo_concurrency.lockutils [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "4fc543a1-34ce-40bb-9e80-72ab9098eb38" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.429246] env[62965]: DEBUG oslo_vmware.api [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5223f2a6-6244-60e2-d090-ba0f53c5cd10, 'name': SearchDatastore_Task, 'duration_secs': 0.010357} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.429872] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.430233] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 856.430578] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.430765] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.432076] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 856.432076] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-50b5e6fa-0099-44f7-bb37-f41f9de83a35 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.440937] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 856.441206] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 856.442264] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cdbb29f7-afaa-4131-8892-99409ff4cff9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.448960] env[62965]: DEBUG oslo_vmware.api [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Waiting for the task: (returnval){ [ 856.448960] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52635e7c-6a3f-267e-2bee-9f7ac1924b42" [ 856.448960] env[62965]: _type = "Task" [ 856.448960] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.457655] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.327s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.458155] env[62965]: DEBUG nova.compute.manager [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 856.461026] env[62965]: DEBUG oslo_vmware.api [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52635e7c-6a3f-267e-2bee-9f7ac1924b42, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.461443] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.440s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.462941] env[62965]: INFO nova.compute.claims [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 856.609914] env[62965]: DEBUG oslo_concurrency.lockutils [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.780957] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389820, 'name': CreateVM_Task, 'duration_secs': 0.29945} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.781393] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 856.781957] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.782317] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.782761] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 856.783151] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a61616fd-a891-404e-92ac-e8020e486362 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.788436] env[62965]: DEBUG oslo_vmware.api [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Waiting for the task: (returnval){ [ 856.788436] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]528ee030-21f6-f82a-f86e-0bdcf7fdbbc2" [ 856.788436] env[62965]: _type = "Task" [ 856.788436] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.796349] env[62965]: DEBUG oslo_vmware.api [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]528ee030-21f6-f82a-f86e-0bdcf7fdbbc2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.815119] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 856.815119] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 856.861137] env[62965]: DEBUG oslo_concurrency.lockutils [req-672ad77f-7f36-4617-8670-48c5f56e8b13 req-035a6ee2-b8a1-423d-a9ea-f267f2a8bf7a service nova] Releasing lock "refresh_cache-64cf898c-2c97-4daa-80c8-0a1c252cbcf2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.960543] env[62965]: DEBUG oslo_vmware.api [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52635e7c-6a3f-267e-2bee-9f7ac1924b42, 'name': SearchDatastore_Task, 'duration_secs': 0.00779} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.961931] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-21d1d72f-d274-44e9-8d4f-0a2d69e527e9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.971159] env[62965]: DEBUG nova.compute.utils [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 856.973825] env[62965]: DEBUG nova.compute.manager [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 856.974342] env[62965]: DEBUG nova.network.neutron [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 856.980047] env[62965]: DEBUG oslo_vmware.api [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Waiting for the task: (returnval){ [ 856.980047] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5213413d-066f-caf0-49bd-70c0c97acbf0" [ 856.980047] env[62965]: _type = "Task" [ 856.980047] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.991659] env[62965]: DEBUG oslo_vmware.api [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5213413d-066f-caf0-49bd-70c0c97acbf0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.058135] env[62965]: DEBUG nova.policy [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e9d9f79c936143f5815fcb99755bbd4e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8fab418e2fe1420793517663574b43bb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 857.300548] env[62965]: DEBUG oslo_vmware.api [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]528ee030-21f6-f82a-f86e-0bdcf7fdbbc2, 'name': SearchDatastore_Task, 'duration_secs': 0.008876} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.301364] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.302122] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 857.302122] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.324337] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 857.324337] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Starting heal instance info cache {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10257}} [ 857.324337] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Rebuilding the list of instances to heal {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10261}} [ 857.475220] env[62965]: DEBUG nova.compute.manager [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 857.492518] env[62965]: DEBUG oslo_vmware.api [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5213413d-066f-caf0-49bd-70c0c97acbf0, 'name': SearchDatastore_Task, 'duration_secs': 0.009067} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.493154] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.493424] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] 64cf898c-2c97-4daa-80c8-0a1c252cbcf2/64cf898c-2c97-4daa-80c8-0a1c252cbcf2.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 857.493697] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.493881] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 857.494179] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-69c01506-8816-434f-8316-bb403ffd228e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.496143] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d001bb2b-b3a8-4b7b-bb86-f74f7fd29440 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.503709] env[62965]: DEBUG oslo_vmware.api [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Waiting for the task: (returnval){ [ 857.503709] env[62965]: value = "task-1389821" [ 857.503709] env[62965]: _type = "Task" [ 857.503709] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.511759] env[62965]: DEBUG oslo_vmware.api [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1389821, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.587603] env[62965]: DEBUG nova.network.neutron [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Successfully created port: 57672e8b-1df1-463d-b9b3-8290667d9fd2 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 857.676655] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-902fff8e-0689-4d08-908f-34d8d6494580 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.684298] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64851aad-4076-4023-9e84-df7925e054c8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.722483] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5eac76d-753d-471f-807b-2fddf90e3862 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.730077] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6abb27d-21d8-4602-b37d-fcb4e2e162cc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.743439] env[62965]: DEBUG nova.compute.provider_tree [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 857.829667] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Skipping network cache update for instance because it is Building. {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10270}} [ 857.829831] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Skipping network cache update for instance because it is Building. {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10270}} [ 857.829978] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Skipping network cache update for instance because it is Building. {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10270}} [ 857.830099] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Skipping network cache update for instance because it is Building. {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10270}} [ 857.830221] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Skipping network cache update for instance because it is Building. {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10270}} [ 857.830342] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Didn't find any instances for network info cache update. {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10343}} [ 857.832938] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 857.832938] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 857.832938] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 857.832938] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 857.832938] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 857.832938] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 857.833351] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62965) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10876}} [ 857.833351] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 858.015321] env[62965]: DEBUG oslo_vmware.api [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1389821, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.248208] env[62965]: DEBUG nova.scheduler.client.report [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 858.337904] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.427807] env[62965]: DEBUG oslo_concurrency.lockutils [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "d038608c-9264-4fd5-b9ba-0653c98ec0e4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.428054] env[62965]: DEBUG oslo_concurrency.lockutils [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "d038608c-9264-4fd5-b9ba-0653c98ec0e4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.487149] env[62965]: DEBUG nova.compute.manager [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 858.511390] env[62965]: DEBUG nova.virt.hardware [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 858.511626] env[62965]: DEBUG nova.virt.hardware [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 858.511783] env[62965]: DEBUG nova.virt.hardware [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 858.511958] env[62965]: DEBUG nova.virt.hardware [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 858.512115] env[62965]: DEBUG nova.virt.hardware [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 858.512261] env[62965]: DEBUG nova.virt.hardware [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 858.512465] env[62965]: DEBUG nova.virt.hardware [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 858.512621] env[62965]: DEBUG nova.virt.hardware [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 858.512784] env[62965]: DEBUG nova.virt.hardware [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 858.512941] env[62965]: DEBUG nova.virt.hardware [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 858.513133] env[62965]: DEBUG nova.virt.hardware [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 858.513887] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2135f37a-5e3f-4754-852e-1a45335ebb85 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.519545] env[62965]: DEBUG oslo_vmware.api [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1389821, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.530663] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-148e6628-2525-4c5c-9ab3-74cc6da5f3e4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.714133] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 858.714292] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 858.715055] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d66c2366-30be-4f59-b77f-da6d6d23905a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.720614] env[62965]: DEBUG oslo_vmware.api [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Waiting for the task: (returnval){ [ 858.720614] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52134d19-0f81-4e91-8223-dec705d75cd1" [ 858.720614] env[62965]: _type = "Task" [ 858.720614] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.727894] env[62965]: DEBUG oslo_vmware.api [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52134d19-0f81-4e91-8223-dec705d75cd1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.754718] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.293s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.755233] env[62965]: DEBUG nova.compute.manager [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 858.757749] env[62965]: DEBUG oslo_concurrency.lockutils [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.783s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.759230] env[62965]: INFO nova.compute.claims [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 859.017533] env[62965]: DEBUG oslo_vmware.api [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1389821, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.138975] env[62965]: DEBUG nova.compute.manager [req-50c0484f-4cfd-42fd-8217-07d39b031a61 req-dc6362bd-09fd-4358-bf76-b328703ea57c service nova] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Received event network-vif-plugged-57672e8b-1df1-463d-b9b3-8290667d9fd2 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 859.139278] env[62965]: DEBUG oslo_concurrency.lockutils [req-50c0484f-4cfd-42fd-8217-07d39b031a61 req-dc6362bd-09fd-4358-bf76-b328703ea57c service nova] Acquiring lock "c2448b76-1553-4b68-a731-0bb0967f4c1d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.139834] env[62965]: DEBUG oslo_concurrency.lockutils [req-50c0484f-4cfd-42fd-8217-07d39b031a61 req-dc6362bd-09fd-4358-bf76-b328703ea57c service nova] Lock "c2448b76-1553-4b68-a731-0bb0967f4c1d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.139931] env[62965]: DEBUG oslo_concurrency.lockutils [req-50c0484f-4cfd-42fd-8217-07d39b031a61 req-dc6362bd-09fd-4358-bf76-b328703ea57c service nova] Lock "c2448b76-1553-4b68-a731-0bb0967f4c1d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.140178] env[62965]: DEBUG nova.compute.manager [req-50c0484f-4cfd-42fd-8217-07d39b031a61 req-dc6362bd-09fd-4358-bf76-b328703ea57c service nova] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] No waiting events found dispatching network-vif-plugged-57672e8b-1df1-463d-b9b3-8290667d9fd2 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 859.140397] env[62965]: WARNING nova.compute.manager [req-50c0484f-4cfd-42fd-8217-07d39b031a61 req-dc6362bd-09fd-4358-bf76-b328703ea57c service nova] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Received unexpected event network-vif-plugged-57672e8b-1df1-463d-b9b3-8290667d9fd2 for instance with vm_state building and task_state spawning. [ 859.230711] env[62965]: DEBUG oslo_vmware.api [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52134d19-0f81-4e91-8223-dec705d75cd1, 'name': SearchDatastore_Task, 'duration_secs': 0.008468} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.231444] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9dd517d-e851-4fb9-84f3-fc9314e1d96c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.236093] env[62965]: DEBUG oslo_vmware.api [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Waiting for the task: (returnval){ [ 859.236093] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5264470f-0acc-7982-23fc-4aebe950b7ff" [ 859.236093] env[62965]: _type = "Task" [ 859.236093] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.243027] env[62965]: DEBUG oslo_vmware.api [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5264470f-0acc-7982-23fc-4aebe950b7ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.250034] env[62965]: DEBUG nova.network.neutron [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Successfully updated port: 57672e8b-1df1-463d-b9b3-8290667d9fd2 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 859.266077] env[62965]: DEBUG nova.compute.utils [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 859.269923] env[62965]: DEBUG nova.compute.manager [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 859.270127] env[62965]: DEBUG nova.network.neutron [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 859.310494] env[62965]: DEBUG nova.policy [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ce0e7f3429454fd18d54cbb964dfb561', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e1c6e8403409438f9ff10235cd2533bc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 859.520236] env[62965]: DEBUG oslo_vmware.api [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1389821, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.681334} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.520488] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] 64cf898c-2c97-4daa-80c8-0a1c252cbcf2/64cf898c-2c97-4daa-80c8-0a1c252cbcf2.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 859.520699] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 859.520946] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9d767080-197c-4045-99aa-75a52445ea3f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.527322] env[62965]: DEBUG oslo_vmware.api [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Waiting for the task: (returnval){ [ 859.527322] env[62965]: value = "task-1389822" [ 859.527322] env[62965]: _type = "Task" [ 859.527322] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.535908] env[62965]: DEBUG oslo_vmware.api [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1389822, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.674743] env[62965]: DEBUG nova.network.neutron [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Successfully created port: dc435eb8-9509-4115-b425-bbd397d12290 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 859.746552] env[62965]: DEBUG oslo_vmware.api [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5264470f-0acc-7982-23fc-4aebe950b7ff, 'name': SearchDatastore_Task, 'duration_secs': 0.008518} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.746751] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.747031] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c/d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 859.747268] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-053ee7b0-14c8-4279-a1f2-34124d6432b2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.753850] env[62965]: DEBUG oslo_vmware.api [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Waiting for the task: (returnval){ [ 859.753850] env[62965]: value = "task-1389823" [ 859.753850] env[62965]: _type = "Task" [ 859.753850] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.757136] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "refresh_cache-c2448b76-1553-4b68-a731-0bb0967f4c1d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.757301] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquired lock "refresh_cache-c2448b76-1553-4b68-a731-0bb0967f4c1d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.757474] env[62965]: DEBUG nova.network.neutron [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 859.763227] env[62965]: DEBUG oslo_vmware.api [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': task-1389823, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.769469] env[62965]: DEBUG nova.compute.manager [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 859.960376] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da0eee29-ac00-4990-a68c-b46516c0d316 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.971720] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e3a97b2-d1fa-46a0-b89d-c50a94ff93e7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.005279] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d07122f-4d00-4397-9a0d-8b0f90835a8a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.015031] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe1d372d-582a-4aaa-a2c1-b226a5583294 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.030318] env[62965]: DEBUG nova.compute.provider_tree [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 860.042978] env[62965]: DEBUG oslo_vmware.api [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1389822, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058159} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.042978] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 860.043917] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-759d616d-cfa9-4d6a-8934-9892a7f60b21 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.066952] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Reconfiguring VM instance instance-00000040 to attach disk [datastore1] 64cf898c-2c97-4daa-80c8-0a1c252cbcf2/64cf898c-2c97-4daa-80c8-0a1c252cbcf2.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 860.067660] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1fc59aaa-1010-4fc0-94a2-69637e127b6e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.089354] env[62965]: DEBUG oslo_vmware.api [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Waiting for the task: (returnval){ [ 860.089354] env[62965]: value = "task-1389824" [ 860.089354] env[62965]: _type = "Task" [ 860.089354] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.098159] env[62965]: DEBUG oslo_vmware.api [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1389824, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.269243] env[62965]: DEBUG oslo_vmware.api [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': task-1389823, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.295042] env[62965]: DEBUG nova.network.neutron [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 860.485914] env[62965]: DEBUG nova.network.neutron [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Updating instance_info_cache with network_info: [{"id": "57672e8b-1df1-463d-b9b3-8290667d9fd2", "address": "fa:16:3e:53:46:47", "network": {"id": "4649fed7-c2ca-43a9-a183-7c99423da726", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1978698173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fab418e2fe1420793517663574b43bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43ad01d2-c7dd-453c-a929-8ad76294d13c", "external-id": "nsx-vlan-transportzone-176", "segmentation_id": 176, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57672e8b-1d", "ovs_interfaceid": "57672e8b-1df1-463d-b9b3-8290667d9fd2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.536537] env[62965]: DEBUG nova.scheduler.client.report [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 860.600329] env[62965]: DEBUG oslo_vmware.api [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1389824, 'name': ReconfigVM_Task, 'duration_secs': 0.327315} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.600464] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Reconfigured VM instance instance-00000040 to attach disk [datastore1] 64cf898c-2c97-4daa-80c8-0a1c252cbcf2/64cf898c-2c97-4daa-80c8-0a1c252cbcf2.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 860.601152] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-78c3f83d-372a-44e2-a78a-4fb291361478 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.607834] env[62965]: DEBUG oslo_vmware.api [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Waiting for the task: (returnval){ [ 860.607834] env[62965]: value = "task-1389825" [ 860.607834] env[62965]: _type = "Task" [ 860.607834] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.616096] env[62965]: DEBUG oslo_vmware.api [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1389825, 'name': Rename_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.765058] env[62965]: DEBUG oslo_vmware.api [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': task-1389823, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.567459} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.765329] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c/d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 860.765539] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 860.765788] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6e0bb132-1e2c-4f1e-8d35-72aeaeed8c8d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.771631] env[62965]: DEBUG oslo_vmware.api [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Waiting for the task: (returnval){ [ 860.771631] env[62965]: value = "task-1389826" [ 860.771631] env[62965]: _type = "Task" [ 860.771631] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.779719] env[62965]: DEBUG nova.compute.manager [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 860.781660] env[62965]: DEBUG oslo_vmware.api [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': task-1389826, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.806742] env[62965]: DEBUG nova.virt.hardware [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 860.806958] env[62965]: DEBUG nova.virt.hardware [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 860.807066] env[62965]: DEBUG nova.virt.hardware [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 860.807242] env[62965]: DEBUG nova.virt.hardware [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 860.807429] env[62965]: DEBUG nova.virt.hardware [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 860.807523] env[62965]: DEBUG nova.virt.hardware [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 860.807708] env[62965]: DEBUG nova.virt.hardware [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 860.807863] env[62965]: DEBUG nova.virt.hardware [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 860.808057] env[62965]: DEBUG nova.virt.hardware [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 860.808227] env[62965]: DEBUG nova.virt.hardware [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 860.808396] env[62965]: DEBUG nova.virt.hardware [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 860.809267] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2414c858-31b9-4e96-8505-ae58e492393f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.816720] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f19d1b6-0d0e-4a2e-93a1-a0e0298e1b0a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.988646] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Releasing lock "refresh_cache-c2448b76-1553-4b68-a731-0bb0967f4c1d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.988986] env[62965]: DEBUG nova.compute.manager [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Instance network_info: |[{"id": "57672e8b-1df1-463d-b9b3-8290667d9fd2", "address": "fa:16:3e:53:46:47", "network": {"id": "4649fed7-c2ca-43a9-a183-7c99423da726", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1978698173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fab418e2fe1420793517663574b43bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43ad01d2-c7dd-453c-a929-8ad76294d13c", "external-id": "nsx-vlan-transportzone-176", "segmentation_id": 176, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57672e8b-1d", "ovs_interfaceid": "57672e8b-1df1-463d-b9b3-8290667d9fd2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 860.989425] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:53:46:47', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '43ad01d2-c7dd-453c-a929-8ad76294d13c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '57672e8b-1df1-463d-b9b3-8290667d9fd2', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 860.996671] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Creating folder: Project (8fab418e2fe1420793517663574b43bb). Parent ref: group-v294931. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 860.996944] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aacbee3b-26fe-43f1-8803-354a08105412 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.008577] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Created folder: Project (8fab418e2fe1420793517663574b43bb) in parent group-v294931. [ 861.008760] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Creating folder: Instances. Parent ref: group-v294959. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 861.008974] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b12e8a61-8c2f-41cd-b2ac-d500a7726b2e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.018223] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Created folder: Instances in parent group-v294959. [ 861.018493] env[62965]: DEBUG oslo.service.loopingcall [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 861.018598] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 861.018805] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c5e97b7b-882f-48ca-8ad9-82ed554d02df {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.037297] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 861.037297] env[62965]: value = "task-1389829" [ 861.037297] env[62965]: _type = "Task" [ 861.037297] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.041185] env[62965]: DEBUG oslo_concurrency.lockutils [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.283s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.041660] env[62965]: DEBUG nova.compute.manager [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 861.047161] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.337s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.052947] env[62965]: INFO nova.compute.claims [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 861.057554] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389829, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.119608] env[62965]: DEBUG oslo_vmware.api [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1389825, 'name': Rename_Task, 'duration_secs': 0.222578} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.119608] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 861.119608] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6468203a-239b-4617-b28f-8d080016b346 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.126022] env[62965]: DEBUG oslo_vmware.api [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Waiting for the task: (returnval){ [ 861.126022] env[62965]: value = "task-1389830" [ 861.126022] env[62965]: _type = "Task" [ 861.126022] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.132878] env[62965]: DEBUG oslo_vmware.api [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1389830, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.168779] env[62965]: DEBUG nova.compute.manager [req-483fb94f-ef59-4bcc-9360-69e4495aabb2 req-882d18d0-d7e8-43cb-ab4a-34e73c11416e service nova] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Received event network-changed-57672e8b-1df1-463d-b9b3-8290667d9fd2 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 861.168978] env[62965]: DEBUG nova.compute.manager [req-483fb94f-ef59-4bcc-9360-69e4495aabb2 req-882d18d0-d7e8-43cb-ab4a-34e73c11416e service nova] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Refreshing instance network info cache due to event network-changed-57672e8b-1df1-463d-b9b3-8290667d9fd2. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 861.169313] env[62965]: DEBUG oslo_concurrency.lockutils [req-483fb94f-ef59-4bcc-9360-69e4495aabb2 req-882d18d0-d7e8-43cb-ab4a-34e73c11416e service nova] Acquiring lock "refresh_cache-c2448b76-1553-4b68-a731-0bb0967f4c1d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.169461] env[62965]: DEBUG oslo_concurrency.lockutils [req-483fb94f-ef59-4bcc-9360-69e4495aabb2 req-882d18d0-d7e8-43cb-ab4a-34e73c11416e service nova] Acquired lock "refresh_cache-c2448b76-1553-4b68-a731-0bb0967f4c1d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.169711] env[62965]: DEBUG nova.network.neutron [req-483fb94f-ef59-4bcc-9360-69e4495aabb2 req-882d18d0-d7e8-43cb-ab4a-34e73c11416e service nova] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Refreshing network info cache for port 57672e8b-1df1-463d-b9b3-8290667d9fd2 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 861.189859] env[62965]: DEBUG nova.network.neutron [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Successfully updated port: dc435eb8-9509-4115-b425-bbd397d12290 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 861.281253] env[62965]: DEBUG oslo_vmware.api [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': task-1389826, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067505} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.281521] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 861.282407] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-022b3c13-def8-42ca-b6b6-44e11464def0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.301899] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c/d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 861.302607] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9787ee4a-4f64-4ef3-9c44-dad817563cee {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.323772] env[62965]: DEBUG oslo_vmware.api [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Waiting for the task: (returnval){ [ 861.323772] env[62965]: value = "task-1389831" [ 861.323772] env[62965]: _type = "Task" [ 861.323772] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.330493] env[62965]: DEBUG oslo_vmware.api [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': task-1389831, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.547337] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389829, 'name': CreateVM_Task, 'duration_secs': 0.507691} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.547710] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 861.548344] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.548534] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.548872] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 861.549159] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84374320-2aec-4340-ba3d-f339352492a6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.553685] env[62965]: DEBUG oslo_vmware.api [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 861.553685] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52d9c8c2-e755-2e37-b5f7-b32bdf53d23e" [ 861.553685] env[62965]: _type = "Task" [ 861.553685] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.557892] env[62965]: DEBUG nova.compute.utils [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 861.559357] env[62965]: DEBUG nova.compute.manager [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 861.559825] env[62965]: DEBUG nova.network.neutron [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 861.569251] env[62965]: DEBUG oslo_vmware.api [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52d9c8c2-e755-2e37-b5f7-b32bdf53d23e, 'name': SearchDatastore_Task, 'duration_secs': 0.012331} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.569597] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.569822] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 861.570065] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.570210] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.570387] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 861.571530] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-992fbfff-86b7-432f-a9ab-a83cd11a5f93 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.579249] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 861.579431] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 861.580146] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f927df87-eea4-4816-af89-f64336189e6a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.587115] env[62965]: DEBUG oslo_vmware.api [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 861.587115] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52326409-52f0-d577-226c-c87d572c2ff1" [ 861.587115] env[62965]: _type = "Task" [ 861.587115] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.601265] env[62965]: DEBUG oslo_vmware.api [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52326409-52f0-d577-226c-c87d572c2ff1, 'name': SearchDatastore_Task, 'duration_secs': 0.008683} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.602025] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5cb25e23-e6d4-4e6f-bb29-c1f42c53a5f9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.613930] env[62965]: DEBUG oslo_vmware.api [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 861.613930] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]528632cb-5a98-492b-c9e6-41cd920eaf78" [ 861.613930] env[62965]: _type = "Task" [ 861.613930] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.622111] env[62965]: DEBUG oslo_vmware.api [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]528632cb-5a98-492b-c9e6-41cd920eaf78, 'name': SearchDatastore_Task, 'duration_secs': 0.008279} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.622371] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.622627] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] c2448b76-1553-4b68-a731-0bb0967f4c1d/c2448b76-1553-4b68-a731-0bb0967f4c1d.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 861.623119] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5d550f4f-bf82-4b3a-933b-2fe14ad96614 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.633699] env[62965]: DEBUG oslo_vmware.api [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1389830, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.635245] env[62965]: DEBUG oslo_vmware.api [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 861.635245] env[62965]: value = "task-1389832" [ 861.635245] env[62965]: _type = "Task" [ 861.635245] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.642885] env[62965]: DEBUG oslo_vmware.api [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1389832, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.682826] env[62965]: DEBUG nova.policy [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '736eaca6d3584f74856c61ff0e6ce03e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd3edf0643c9d40e9bb568742b7a7a508', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 861.694877] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquiring lock "refresh_cache-d4d9b202-3aa7-4b66-949d-4a2eb34b3adf" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.695039] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquired lock "refresh_cache-d4d9b202-3aa7-4b66-949d-4a2eb34b3adf" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.695187] env[62965]: DEBUG nova.network.neutron [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 861.832353] env[62965]: DEBUG oslo_vmware.api [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': task-1389831, 'name': ReconfigVM_Task, 'duration_secs': 0.4435} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.832690] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Reconfigured VM instance instance-00000041 to attach disk [datastore1] d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c/d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 861.833327] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-15af4565-ea96-42ed-ad89-a3774485e15b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.839970] env[62965]: DEBUG oslo_vmware.api [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Waiting for the task: (returnval){ [ 861.839970] env[62965]: value = "task-1389833" [ 861.839970] env[62965]: _type = "Task" [ 861.839970] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.852084] env[62965]: DEBUG oslo_vmware.api [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': task-1389833, 'name': Rename_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.920673] env[62965]: DEBUG nova.network.neutron [req-483fb94f-ef59-4bcc-9360-69e4495aabb2 req-882d18d0-d7e8-43cb-ab4a-34e73c11416e service nova] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Updated VIF entry in instance network info cache for port 57672e8b-1df1-463d-b9b3-8290667d9fd2. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 861.921225] env[62965]: DEBUG nova.network.neutron [req-483fb94f-ef59-4bcc-9360-69e4495aabb2 req-882d18d0-d7e8-43cb-ab4a-34e73c11416e service nova] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Updating instance_info_cache with network_info: [{"id": "57672e8b-1df1-463d-b9b3-8290667d9fd2", "address": "fa:16:3e:53:46:47", "network": {"id": "4649fed7-c2ca-43a9-a183-7c99423da726", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1978698173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fab418e2fe1420793517663574b43bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43ad01d2-c7dd-453c-a929-8ad76294d13c", "external-id": "nsx-vlan-transportzone-176", "segmentation_id": 176, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57672e8b-1d", "ovs_interfaceid": "57672e8b-1df1-463d-b9b3-8290667d9fd2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.064868] env[62965]: DEBUG nova.compute.manager [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 862.073065] env[62965]: DEBUG nova.network.neutron [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Successfully created port: 2cf58a6d-2155-49a4-a049-98ef841cc561 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 862.139308] env[62965]: DEBUG oslo_vmware.api [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1389830, 'name': PowerOnVM_Task, 'duration_secs': 0.613564} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.144303] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 862.144303] env[62965]: INFO nova.compute.manager [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Took 10.08 seconds to spawn the instance on the hypervisor. [ 862.144303] env[62965]: DEBUG nova.compute.manager [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 862.144303] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df5a1c7b-2aea-45e7-9d91-b606cc401634 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.160708] env[62965]: DEBUG oslo_vmware.api [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1389832, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.241139] env[62965]: DEBUG nova.network.neutron [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 862.307771] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b1bcf64-51f8-4966-8e77-ba78881c7bbe {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.318958] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b63caf58-487c-49a6-aa30-7268591dfd3b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.365027] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bec88c5-1658-4a74-83b7-1d81ca0b1e89 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.372342] env[62965]: DEBUG oslo_vmware.api [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': task-1389833, 'name': Rename_Task, 'duration_secs': 0.459813} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.374296] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 862.374508] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1054cc1c-009f-4310-9e54-97987b3514bd {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.376722] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9873bf7-89c4-405e-9d73-90b0fb3f5bd7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.389738] env[62965]: DEBUG nova.compute.provider_tree [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 862.392815] env[62965]: DEBUG oslo_vmware.api [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Waiting for the task: (returnval){ [ 862.392815] env[62965]: value = "task-1389834" [ 862.392815] env[62965]: _type = "Task" [ 862.392815] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.399126] env[62965]: DEBUG oslo_vmware.api [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': task-1389834, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.401015] env[62965]: DEBUG nova.network.neutron [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Updating instance_info_cache with network_info: [{"id": "dc435eb8-9509-4115-b425-bbd397d12290", "address": "fa:16:3e:b6:35:48", "network": {"id": "51f79fa0-8980-40d2-9255-a8da5b8fc5d7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-909239894-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1c6e8403409438f9ff10235cd2533bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc435eb8-95", "ovs_interfaceid": "dc435eb8-9509-4115-b425-bbd397d12290", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.424049] env[62965]: DEBUG oslo_concurrency.lockutils [req-483fb94f-ef59-4bcc-9360-69e4495aabb2 req-882d18d0-d7e8-43cb-ab4a-34e73c11416e service nova] Releasing lock "refresh_cache-c2448b76-1553-4b68-a731-0bb0967f4c1d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.424330] env[62965]: DEBUG nova.compute.manager [req-483fb94f-ef59-4bcc-9360-69e4495aabb2 req-882d18d0-d7e8-43cb-ab4a-34e73c11416e service nova] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Received event network-vif-plugged-dc435eb8-9509-4115-b425-bbd397d12290 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 862.424524] env[62965]: DEBUG oslo_concurrency.lockutils [req-483fb94f-ef59-4bcc-9360-69e4495aabb2 req-882d18d0-d7e8-43cb-ab4a-34e73c11416e service nova] Acquiring lock "d4d9b202-3aa7-4b66-949d-4a2eb34b3adf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.424727] env[62965]: DEBUG oslo_concurrency.lockutils [req-483fb94f-ef59-4bcc-9360-69e4495aabb2 req-882d18d0-d7e8-43cb-ab4a-34e73c11416e service nova] Lock "d4d9b202-3aa7-4b66-949d-4a2eb34b3adf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.424884] env[62965]: DEBUG oslo_concurrency.lockutils [req-483fb94f-ef59-4bcc-9360-69e4495aabb2 req-882d18d0-d7e8-43cb-ab4a-34e73c11416e service nova] Lock "d4d9b202-3aa7-4b66-949d-4a2eb34b3adf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.425058] env[62965]: DEBUG nova.compute.manager [req-483fb94f-ef59-4bcc-9360-69e4495aabb2 req-882d18d0-d7e8-43cb-ab4a-34e73c11416e service nova] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] No waiting events found dispatching network-vif-plugged-dc435eb8-9509-4115-b425-bbd397d12290 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 862.425483] env[62965]: WARNING nova.compute.manager [req-483fb94f-ef59-4bcc-9360-69e4495aabb2 req-882d18d0-d7e8-43cb-ab4a-34e73c11416e service nova] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Received unexpected event network-vif-plugged-dc435eb8-9509-4115-b425-bbd397d12290 for instance with vm_state building and task_state spawning. [ 862.650028] env[62965]: DEBUG oslo_vmware.api [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1389832, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.579206} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.650318] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] c2448b76-1553-4b68-a731-0bb0967f4c1d/c2448b76-1553-4b68-a731-0bb0967f4c1d.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 862.650546] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 862.650823] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-132afdc9-d133-4432-8506-dddde8fc7f09 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.657550] env[62965]: DEBUG oslo_vmware.api [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 862.657550] env[62965]: value = "task-1389835" [ 862.657550] env[62965]: _type = "Task" [ 862.657550] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.672109] env[62965]: DEBUG oslo_vmware.api [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1389835, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.676788] env[62965]: INFO nova.compute.manager [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Took 29.51 seconds to build instance. [ 862.897560] env[62965]: DEBUG nova.scheduler.client.report [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 862.903031] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Releasing lock "refresh_cache-d4d9b202-3aa7-4b66-949d-4a2eb34b3adf" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.903168] env[62965]: DEBUG nova.compute.manager [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Instance network_info: |[{"id": "dc435eb8-9509-4115-b425-bbd397d12290", "address": "fa:16:3e:b6:35:48", "network": {"id": "51f79fa0-8980-40d2-9255-a8da5b8fc5d7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-909239894-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1c6e8403409438f9ff10235cd2533bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc435eb8-95", "ovs_interfaceid": "dc435eb8-9509-4115-b425-bbd397d12290", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 862.906260] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b6:35:48', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c6934071-bf85-4591-9c7d-55c7ea131262', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dc435eb8-9509-4115-b425-bbd397d12290', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 862.914148] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Creating folder: Project (e1c6e8403409438f9ff10235cd2533bc). Parent ref: group-v294931. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 862.914816] env[62965]: DEBUG oslo_vmware.api [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': task-1389834, 'name': PowerOnVM_Task, 'duration_secs': 0.445161} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.915196] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-823620d6-9719-4c0e-a18d-8804fd5aea34 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.916842] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 862.917063] env[62965]: INFO nova.compute.manager [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Took 6.76 seconds to spawn the instance on the hypervisor. [ 862.917242] env[62965]: DEBUG nova.compute.manager [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 862.918262] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5238c99e-df7e-40c9-b112-04a9948cc2fb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.930572] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Created folder: Project (e1c6e8403409438f9ff10235cd2533bc) in parent group-v294931. [ 862.930753] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Creating folder: Instances. Parent ref: group-v294962. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 862.930956] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-beafe8b1-aaec-497c-8a70-91459ece8d75 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.939911] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Created folder: Instances in parent group-v294962. [ 862.939911] env[62965]: DEBUG oslo.service.loopingcall [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 862.939911] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 862.939911] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1ec40687-f2ad-4f7a-af3a-516825305609 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.962870] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 862.962870] env[62965]: value = "task-1389838" [ 862.962870] env[62965]: _type = "Task" [ 862.962870] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.971472] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389838, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.077171] env[62965]: DEBUG nova.compute.manager [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 863.103108] env[62965]: DEBUG nova.virt.hardware [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 863.103359] env[62965]: DEBUG nova.virt.hardware [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 863.103511] env[62965]: DEBUG nova.virt.hardware [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 863.103688] env[62965]: DEBUG nova.virt.hardware [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 863.103832] env[62965]: DEBUG nova.virt.hardware [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 863.103996] env[62965]: DEBUG nova.virt.hardware [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 863.104509] env[62965]: DEBUG nova.virt.hardware [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 863.104706] env[62965]: DEBUG nova.virt.hardware [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 863.104906] env[62965]: DEBUG nova.virt.hardware [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 863.105117] env[62965]: DEBUG nova.virt.hardware [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 863.105303] env[62965]: DEBUG nova.virt.hardware [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 863.106331] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1ad0851-8e46-4510-a4f2-dc8906122929 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.114691] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72958bc8-e667-489b-81bb-aa99e5ece225 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.167793] env[62965]: DEBUG oslo_vmware.api [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1389835, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072926} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.168111] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 863.168886] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7bda716-1b0e-43b7-9665-ed0d5e09e792 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.183017] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ca38dc11-2a44-4019-b9af-0a3ab7a16bbb tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Lock "64cf898c-2c97-4daa-80c8-0a1c252cbcf2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.068s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.191910] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] c2448b76-1553-4b68-a731-0bb0967f4c1d/c2448b76-1553-4b68-a731-0bb0967f4c1d.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 863.193235] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2ab19e7f-7541-4fb9-b13e-0a102dc6d0c3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.220151] env[62965]: DEBUG nova.compute.manager [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 863.222914] env[62965]: DEBUG nova.compute.manager [req-63818919-6769-4d38-9a0d-a139eb4bb83e req-7648df14-6ad3-4373-a0ba-52439bb5fa14 service nova] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Received event network-changed-dc435eb8-9509-4115-b425-bbd397d12290 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 863.223130] env[62965]: DEBUG nova.compute.manager [req-63818919-6769-4d38-9a0d-a139eb4bb83e req-7648df14-6ad3-4373-a0ba-52439bb5fa14 service nova] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Refreshing instance network info cache due to event network-changed-dc435eb8-9509-4115-b425-bbd397d12290. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 863.223363] env[62965]: DEBUG oslo_concurrency.lockutils [req-63818919-6769-4d38-9a0d-a139eb4bb83e req-7648df14-6ad3-4373-a0ba-52439bb5fa14 service nova] Acquiring lock "refresh_cache-d4d9b202-3aa7-4b66-949d-4a2eb34b3adf" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.223508] env[62965]: DEBUG oslo_concurrency.lockutils [req-63818919-6769-4d38-9a0d-a139eb4bb83e req-7648df14-6ad3-4373-a0ba-52439bb5fa14 service nova] Acquired lock "refresh_cache-d4d9b202-3aa7-4b66-949d-4a2eb34b3adf" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.223665] env[62965]: DEBUG nova.network.neutron [req-63818919-6769-4d38-9a0d-a139eb4bb83e req-7648df14-6ad3-4373-a0ba-52439bb5fa14 service nova] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Refreshing network info cache for port dc435eb8-9509-4115-b425-bbd397d12290 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 863.234666] env[62965]: DEBUG oslo_vmware.api [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 863.234666] env[62965]: value = "task-1389839" [ 863.234666] env[62965]: _type = "Task" [ 863.234666] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.247232] env[62965]: DEBUG oslo_vmware.api [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1389839, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.403337] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.356s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.404172] env[62965]: DEBUG nova.compute.manager [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 863.408177] env[62965]: DEBUG oslo_concurrency.lockutils [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.035s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.412129] env[62965]: INFO nova.compute.claims [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 863.438530] env[62965]: INFO nova.compute.manager [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Took 26.05 seconds to build instance. [ 863.479031] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389838, 'name': CreateVM_Task, 'duration_secs': 0.362942} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.479031] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 863.479031] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.479031] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.479031] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 863.479031] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1497cfd-4f02-4f65-91a0-e50e1950bb1f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.483940] env[62965]: DEBUG oslo_vmware.api [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 863.483940] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52f72a71-3747-984e-d098-5452242e2098" [ 863.483940] env[62965]: _type = "Task" [ 863.483940] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.490825] env[62965]: DEBUG oslo_vmware.api [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52f72a71-3747-984e-d098-5452242e2098, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.706137] env[62965]: DEBUG nova.compute.manager [req-54d989d8-65cd-4d8c-b59d-54bb4e854441 req-b5dff26a-39cf-4843-9833-d37fc085bb26 service nova] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Received event network-changed-97930362-17f2-40a5-bc09-a9df5f010935 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 863.706404] env[62965]: DEBUG nova.compute.manager [req-54d989d8-65cd-4d8c-b59d-54bb4e854441 req-b5dff26a-39cf-4843-9833-d37fc085bb26 service nova] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Refreshing instance network info cache due to event network-changed-97930362-17f2-40a5-bc09-a9df5f010935. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 863.706649] env[62965]: DEBUG oslo_concurrency.lockutils [req-54d989d8-65cd-4d8c-b59d-54bb4e854441 req-b5dff26a-39cf-4843-9833-d37fc085bb26 service nova] Acquiring lock "refresh_cache-64cf898c-2c97-4daa-80c8-0a1c252cbcf2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.706749] env[62965]: DEBUG oslo_concurrency.lockutils [req-54d989d8-65cd-4d8c-b59d-54bb4e854441 req-b5dff26a-39cf-4843-9833-d37fc085bb26 service nova] Acquired lock "refresh_cache-64cf898c-2c97-4daa-80c8-0a1c252cbcf2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.706865] env[62965]: DEBUG nova.network.neutron [req-54d989d8-65cd-4d8c-b59d-54bb4e854441 req-b5dff26a-39cf-4843-9833-d37fc085bb26 service nova] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Refreshing network info cache for port 97930362-17f2-40a5-bc09-a9df5f010935 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 863.750361] env[62965]: DEBUG oslo_vmware.api [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1389839, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.759796] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.767880] env[62965]: DEBUG nova.network.neutron [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Successfully updated port: 2cf58a6d-2155-49a4-a049-98ef841cc561 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 863.917106] env[62965]: DEBUG nova.compute.utils [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 863.920775] env[62965]: DEBUG nova.compute.manager [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 863.920967] env[62965]: DEBUG nova.network.neutron [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 863.941047] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ccca0c24-2b41-4536-a26c-85848469804a tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Lock "d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.357s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.965594] env[62965]: DEBUG nova.network.neutron [req-63818919-6769-4d38-9a0d-a139eb4bb83e req-7648df14-6ad3-4373-a0ba-52439bb5fa14 service nova] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Updated VIF entry in instance network info cache for port dc435eb8-9509-4115-b425-bbd397d12290. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 863.965594] env[62965]: DEBUG nova.network.neutron [req-63818919-6769-4d38-9a0d-a139eb4bb83e req-7648df14-6ad3-4373-a0ba-52439bb5fa14 service nova] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Updating instance_info_cache with network_info: [{"id": "dc435eb8-9509-4115-b425-bbd397d12290", "address": "fa:16:3e:b6:35:48", "network": {"id": "51f79fa0-8980-40d2-9255-a8da5b8fc5d7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-909239894-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1c6e8403409438f9ff10235cd2533bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc435eb8-95", "ovs_interfaceid": "dc435eb8-9509-4115-b425-bbd397d12290", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.974226] env[62965]: DEBUG nova.policy [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e935429810f44ac09e07e1ba08916e75', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c1caebff08644e21b3f1c461aa115b8e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 864.000653] env[62965]: DEBUG oslo_vmware.api [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52f72a71-3747-984e-d098-5452242e2098, 'name': SearchDatastore_Task, 'duration_secs': 0.025639} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.002279] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.002683] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 864.004226] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.004226] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.004226] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 864.004226] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e3077612-3391-438b-917d-ddc8cfa934d1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.013421] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 864.013788] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 864.016903] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d125f068-91ac-49c2-9ff8-67812ba5b634 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.022548] env[62965]: DEBUG oslo_vmware.api [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 864.022548] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]525cfa8d-03aa-35c2-25e4-0b86b95385b6" [ 864.022548] env[62965]: _type = "Task" [ 864.022548] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.030898] env[62965]: DEBUG oslo_vmware.api [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]525cfa8d-03aa-35c2-25e4-0b86b95385b6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.247389] env[62965]: DEBUG oslo_vmware.api [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1389839, 'name': ReconfigVM_Task, 'duration_secs': 0.852843} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.248212] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Reconfigured VM instance instance-00000042 to attach disk [datastore1] c2448b76-1553-4b68-a731-0bb0967f4c1d/c2448b76-1553-4b68-a731-0bb0967f4c1d.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 864.248212] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-360ae6bd-5835-4f8f-acaa-b1a3db3014bb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.257505] env[62965]: DEBUG oslo_vmware.api [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 864.257505] env[62965]: value = "task-1389840" [ 864.257505] env[62965]: _type = "Task" [ 864.257505] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.272022] env[62965]: DEBUG oslo_concurrency.lockutils [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Acquiring lock "refresh_cache-b0d8f38a-86d3-466a-9b83-35145071dc21" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.272137] env[62965]: DEBUG oslo_concurrency.lockutils [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Acquired lock "refresh_cache-b0d8f38a-86d3-466a-9b83-35145071dc21" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.273378] env[62965]: DEBUG nova.network.neutron [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 864.274497] env[62965]: DEBUG oslo_vmware.api [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1389840, 'name': Rename_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.377951] env[62965]: DEBUG nova.network.neutron [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Successfully created port: 80afc0d2-0d83-4374-a4c0-792eabf66c40 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 864.421517] env[62965]: DEBUG nova.compute.manager [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 864.446959] env[62965]: DEBUG nova.compute.manager [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 864.471602] env[62965]: DEBUG oslo_concurrency.lockutils [req-63818919-6769-4d38-9a0d-a139eb4bb83e req-7648df14-6ad3-4373-a0ba-52439bb5fa14 service nova] Releasing lock "refresh_cache-d4d9b202-3aa7-4b66-949d-4a2eb34b3adf" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.474510] env[62965]: DEBUG nova.network.neutron [req-54d989d8-65cd-4d8c-b59d-54bb4e854441 req-b5dff26a-39cf-4843-9833-d37fc085bb26 service nova] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Updated VIF entry in instance network info cache for port 97930362-17f2-40a5-bc09-a9df5f010935. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 864.474913] env[62965]: DEBUG nova.network.neutron [req-54d989d8-65cd-4d8c-b59d-54bb4e854441 req-b5dff26a-39cf-4843-9833-d37fc085bb26 service nova] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Updating instance_info_cache with network_info: [{"id": "97930362-17f2-40a5-bc09-a9df5f010935", "address": "fa:16:3e:a4:ce:72", "network": {"id": "88959f71-7c4c-4c1d-aef7-b674c749fff6", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-692265432-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d9897f889b854681909285c643d1ae86", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2d2742ba-c3af-4412-877d-c2811dfeba46", "external-id": "nsx-vlan-transportzone-390", "segmentation_id": 390, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97930362-17", "ovs_interfaceid": "97930362-17f2-40a5-bc09-a9df5f010935", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.495624] env[62965]: INFO nova.compute.manager [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Rebuilding instance [ 864.536285] env[62965]: DEBUG oslo_vmware.api [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]525cfa8d-03aa-35c2-25e4-0b86b95385b6, 'name': SearchDatastore_Task, 'duration_secs': 0.009712} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.541549] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d66721c3-edc7-47d8-ace3-cbcc1ef328f5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.547754] env[62965]: DEBUG nova.compute.manager [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 864.548954] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6f85f16-1a82-4b0a-986e-56d5535707e5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.554804] env[62965]: DEBUG oslo_vmware.api [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 864.554804] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52dfc505-20b5-b8a6-795c-a2498065671c" [ 864.554804] env[62965]: _type = "Task" [ 864.554804] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.567204] env[62965]: DEBUG oslo_vmware.api [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52dfc505-20b5-b8a6-795c-a2498065671c, 'name': SearchDatastore_Task, 'duration_secs': 0.012512} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.569391] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.569693] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] d4d9b202-3aa7-4b66-949d-4a2eb34b3adf/d4d9b202-3aa7-4b66-949d-4a2eb34b3adf.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 864.570107] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fded5d2c-b4c6-49ed-9af3-d3dc26671ca9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.576636] env[62965]: DEBUG oslo_vmware.api [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 864.576636] env[62965]: value = "task-1389841" [ 864.576636] env[62965]: _type = "Task" [ 864.576636] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.585966] env[62965]: DEBUG oslo_vmware.api [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389841, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.671682] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eafe5861-cad0-409a-b76a-b7550a396708 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.678860] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-586ebdc9-c1ab-4af9-9994-51ee2cb4d968 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.709070] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd265ad9-3d2c-4daf-8063-3e7e68e6bae2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.717359] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-979c3633-fba1-423a-a95b-9bc0cee9fff1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.732229] env[62965]: DEBUG nova.compute.provider_tree [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 864.766951] env[62965]: DEBUG oslo_vmware.api [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1389840, 'name': Rename_Task, 'duration_secs': 0.145896} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.767224] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 864.767478] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-958f72ef-364a-4911-9d8e-df78e1ce9985 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.774136] env[62965]: DEBUG oslo_vmware.api [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 864.774136] env[62965]: value = "task-1389842" [ 864.774136] env[62965]: _type = "Task" [ 864.774136] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.781847] env[62965]: DEBUG oslo_vmware.api [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1389842, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.811548] env[62965]: DEBUG nova.network.neutron [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 864.945208] env[62965]: DEBUG nova.network.neutron [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Updating instance_info_cache with network_info: [{"id": "2cf58a6d-2155-49a4-a049-98ef841cc561", "address": "fa:16:3e:5c:f9:90", "network": {"id": "a27e0199-89db-4b64-b3fd-4a12cd598e60", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-12171013-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d3edf0643c9d40e9bb568742b7a7a508", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f4a795c-8718-4a7c-aafe-9da231df10f8", "external-id": "nsx-vlan-transportzone-162", "segmentation_id": 162, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2cf58a6d-21", "ovs_interfaceid": "2cf58a6d-2155-49a4-a049-98ef841cc561", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.974037] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.976911] env[62965]: DEBUG oslo_concurrency.lockutils [req-54d989d8-65cd-4d8c-b59d-54bb4e854441 req-b5dff26a-39cf-4843-9833-d37fc085bb26 service nova] Releasing lock "refresh_cache-64cf898c-2c97-4daa-80c8-0a1c252cbcf2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.086821] env[62965]: DEBUG oslo_vmware.api [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389841, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.226781] env[62965]: DEBUG nova.compute.manager [req-079d143f-b877-4aa5-8e41-6133f86e93e9 req-d8146b69-973b-4ae0-bc7e-7fb304b3299d service nova] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Received event network-vif-plugged-2cf58a6d-2155-49a4-a049-98ef841cc561 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 865.226979] env[62965]: DEBUG oslo_concurrency.lockutils [req-079d143f-b877-4aa5-8e41-6133f86e93e9 req-d8146b69-973b-4ae0-bc7e-7fb304b3299d service nova] Acquiring lock "b0d8f38a-86d3-466a-9b83-35145071dc21-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.227215] env[62965]: DEBUG oslo_concurrency.lockutils [req-079d143f-b877-4aa5-8e41-6133f86e93e9 req-d8146b69-973b-4ae0-bc7e-7fb304b3299d service nova] Lock "b0d8f38a-86d3-466a-9b83-35145071dc21-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.227453] env[62965]: DEBUG oslo_concurrency.lockutils [req-079d143f-b877-4aa5-8e41-6133f86e93e9 req-d8146b69-973b-4ae0-bc7e-7fb304b3299d service nova] Lock "b0d8f38a-86d3-466a-9b83-35145071dc21-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.227542] env[62965]: DEBUG nova.compute.manager [req-079d143f-b877-4aa5-8e41-6133f86e93e9 req-d8146b69-973b-4ae0-bc7e-7fb304b3299d service nova] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] No waiting events found dispatching network-vif-plugged-2cf58a6d-2155-49a4-a049-98ef841cc561 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 865.227707] env[62965]: WARNING nova.compute.manager [req-079d143f-b877-4aa5-8e41-6133f86e93e9 req-d8146b69-973b-4ae0-bc7e-7fb304b3299d service nova] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Received unexpected event network-vif-plugged-2cf58a6d-2155-49a4-a049-98ef841cc561 for instance with vm_state building and task_state spawning. [ 865.227864] env[62965]: DEBUG nova.compute.manager [req-079d143f-b877-4aa5-8e41-6133f86e93e9 req-d8146b69-973b-4ae0-bc7e-7fb304b3299d service nova] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Received event network-changed-2cf58a6d-2155-49a4-a049-98ef841cc561 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 865.228026] env[62965]: DEBUG nova.compute.manager [req-079d143f-b877-4aa5-8e41-6133f86e93e9 req-d8146b69-973b-4ae0-bc7e-7fb304b3299d service nova] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Refreshing instance network info cache due to event network-changed-2cf58a6d-2155-49a4-a049-98ef841cc561. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 865.228195] env[62965]: DEBUG oslo_concurrency.lockutils [req-079d143f-b877-4aa5-8e41-6133f86e93e9 req-d8146b69-973b-4ae0-bc7e-7fb304b3299d service nova] Acquiring lock "refresh_cache-b0d8f38a-86d3-466a-9b83-35145071dc21" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.234933] env[62965]: DEBUG nova.scheduler.client.report [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 865.286072] env[62965]: DEBUG oslo_vmware.api [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1389842, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.437195] env[62965]: DEBUG nova.compute.manager [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 865.449299] env[62965]: DEBUG oslo_concurrency.lockutils [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Releasing lock "refresh_cache-b0d8f38a-86d3-466a-9b83-35145071dc21" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.449724] env[62965]: DEBUG nova.compute.manager [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Instance network_info: |[{"id": "2cf58a6d-2155-49a4-a049-98ef841cc561", "address": "fa:16:3e:5c:f9:90", "network": {"id": "a27e0199-89db-4b64-b3fd-4a12cd598e60", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-12171013-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d3edf0643c9d40e9bb568742b7a7a508", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f4a795c-8718-4a7c-aafe-9da231df10f8", "external-id": "nsx-vlan-transportzone-162", "segmentation_id": 162, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2cf58a6d-21", "ovs_interfaceid": "2cf58a6d-2155-49a4-a049-98ef841cc561", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 865.450173] env[62965]: DEBUG oslo_concurrency.lockutils [req-079d143f-b877-4aa5-8e41-6133f86e93e9 req-d8146b69-973b-4ae0-bc7e-7fb304b3299d service nova] Acquired lock "refresh_cache-b0d8f38a-86d3-466a-9b83-35145071dc21" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.450445] env[62965]: DEBUG nova.network.neutron [req-079d143f-b877-4aa5-8e41-6133f86e93e9 req-d8146b69-973b-4ae0-bc7e-7fb304b3299d service nova] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Refreshing network info cache for port 2cf58a6d-2155-49a4-a049-98ef841cc561 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 865.456296] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5c:f9:90', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3f4a795c-8718-4a7c-aafe-9da231df10f8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2cf58a6d-2155-49a4-a049-98ef841cc561', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 865.465897] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Creating folder: Project (d3edf0643c9d40e9bb568742b7a7a508). Parent ref: group-v294931. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 865.468962] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c06db28e-e756-495e-a4a4-009fa1d80735 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.480182] env[62965]: DEBUG nova.virt.hardware [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 865.480439] env[62965]: DEBUG nova.virt.hardware [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 865.480616] env[62965]: DEBUG nova.virt.hardware [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 865.480839] env[62965]: DEBUG nova.virt.hardware [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 865.480986] env[62965]: DEBUG nova.virt.hardware [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 865.481148] env[62965]: DEBUG nova.virt.hardware [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 865.481357] env[62965]: DEBUG nova.virt.hardware [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 865.481517] env[62965]: DEBUG nova.virt.hardware [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 865.481685] env[62965]: DEBUG nova.virt.hardware [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 865.481881] env[62965]: DEBUG nova.virt.hardware [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 865.482028] env[62965]: DEBUG nova.virt.hardware [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 865.483270] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ee1739c-e517-445c-a374-54ada46b30ea {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.487888] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Created folder: Project (d3edf0643c9d40e9bb568742b7a7a508) in parent group-v294931. [ 865.488086] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Creating folder: Instances. Parent ref: group-v294965. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 865.488759] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8dc8c28c-b26f-4825-8698-9ad5aaff338a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.497177] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4035203-56cf-4d48-b073-b29d900098d0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.503843] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Created folder: Instances in parent group-v294965. [ 865.504221] env[62965]: DEBUG oslo.service.loopingcall [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 865.505157] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 865.505466] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-833d8ed2-a126-49eb-893a-7c882b2a86b3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.542352] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 865.542352] env[62965]: value = "task-1389845" [ 865.542352] env[62965]: _type = "Task" [ 865.542352] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.550402] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389845, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.566255] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 865.566594] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d8fce26d-ed0b-411f-b8f4-475ec0fa2e6f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.574696] env[62965]: DEBUG oslo_vmware.api [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Waiting for the task: (returnval){ [ 865.574696] env[62965]: value = "task-1389846" [ 865.574696] env[62965]: _type = "Task" [ 865.574696] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.588126] env[62965]: DEBUG oslo_vmware.api [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': task-1389846, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.591440] env[62965]: DEBUG oslo_vmware.api [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389841, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.956921} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.591688] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] d4d9b202-3aa7-4b66-949d-4a2eb34b3adf/d4d9b202-3aa7-4b66-949d-4a2eb34b3adf.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 865.591941] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 865.592210] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d74dc0fd-b115-49bb-bbfb-855674a04181 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.598223] env[62965]: DEBUG oslo_vmware.api [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 865.598223] env[62965]: value = "task-1389847" [ 865.598223] env[62965]: _type = "Task" [ 865.598223] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.606857] env[62965]: DEBUG oslo_vmware.api [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389847, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.740829] env[62965]: DEBUG oslo_concurrency.lockutils [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.333s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.741401] env[62965]: DEBUG nova.compute.manager [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 865.744104] env[62965]: DEBUG oslo_concurrency.lockutils [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.611s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.745579] env[62965]: INFO nova.compute.claims [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 865.788983] env[62965]: DEBUG oslo_vmware.api [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1389842, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.040390] env[62965]: DEBUG nova.network.neutron [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Successfully updated port: 80afc0d2-0d83-4374-a4c0-792eabf66c40 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 866.052840] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389845, 'name': CreateVM_Task, 'duration_secs': 0.395236} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.053640] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 866.054359] env[62965]: DEBUG oslo_concurrency.lockutils [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.055022] env[62965]: DEBUG oslo_concurrency.lockutils [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.055022] env[62965]: DEBUG oslo_concurrency.lockutils [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 866.055377] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed78bbaa-8f7e-4ae0-9cfb-92a25cb4c458 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.060285] env[62965]: DEBUG oslo_vmware.api [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 866.060285] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]527150bc-94d2-07b7-214c-826588c6dfbd" [ 866.060285] env[62965]: _type = "Task" [ 866.060285] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.074395] env[62965]: DEBUG oslo_vmware.api [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]527150bc-94d2-07b7-214c-826588c6dfbd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.088025] env[62965]: DEBUG oslo_vmware.api [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': task-1389846, 'name': PowerOffVM_Task, 'duration_secs': 0.154701} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.088025] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 866.088025] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 866.088025] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f42b3239-0b0e-419f-b569-65f6089a586e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.096155] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 866.096155] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-226fc64a-08cc-498d-a601-cba76d38e042 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.108475] env[62965]: DEBUG oslo_vmware.api [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389847, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.094823} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.109197] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 866.112265] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-524fa1a0-34ae-4daa-85ae-6aa0cbe56d5a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.140285] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] d4d9b202-3aa7-4b66-949d-4a2eb34b3adf/d4d9b202-3aa7-4b66-949d-4a2eb34b3adf.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 866.140464] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 866.141042] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Deleting contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 866.141042] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Deleting the datastore file [datastore1] d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 866.141400] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6fcbf04d-dc6c-4e17-982a-a6fb967e8cd7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.156345] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b554ef51-5645-400a-8e09-33616e05adcf {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.165511] env[62965]: DEBUG oslo_vmware.api [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Waiting for the task: (returnval){ [ 866.165511] env[62965]: value = "task-1389850" [ 866.165511] env[62965]: _type = "Task" [ 866.165511] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.165785] env[62965]: DEBUG oslo_vmware.api [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 866.165785] env[62965]: value = "task-1389849" [ 866.165785] env[62965]: _type = "Task" [ 866.165785] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.178933] env[62965]: DEBUG oslo_vmware.api [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': task-1389850, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.182646] env[62965]: DEBUG oslo_vmware.api [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389849, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.237062] env[62965]: DEBUG nova.network.neutron [req-079d143f-b877-4aa5-8e41-6133f86e93e9 req-d8146b69-973b-4ae0-bc7e-7fb304b3299d service nova] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Updated VIF entry in instance network info cache for port 2cf58a6d-2155-49a4-a049-98ef841cc561. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 866.237512] env[62965]: DEBUG nova.network.neutron [req-079d143f-b877-4aa5-8e41-6133f86e93e9 req-d8146b69-973b-4ae0-bc7e-7fb304b3299d service nova] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Updating instance_info_cache with network_info: [{"id": "2cf58a6d-2155-49a4-a049-98ef841cc561", "address": "fa:16:3e:5c:f9:90", "network": {"id": "a27e0199-89db-4b64-b3fd-4a12cd598e60", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-12171013-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d3edf0643c9d40e9bb568742b7a7a508", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f4a795c-8718-4a7c-aafe-9da231df10f8", "external-id": "nsx-vlan-transportzone-162", "segmentation_id": 162, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2cf58a6d-21", "ovs_interfaceid": "2cf58a6d-2155-49a4-a049-98ef841cc561", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.250319] env[62965]: DEBUG nova.compute.utils [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 866.253937] env[62965]: DEBUG nova.compute.manager [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 866.254156] env[62965]: DEBUG nova.network.neutron [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 866.290251] env[62965]: DEBUG oslo_vmware.api [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1389842, 'name': PowerOnVM_Task, 'duration_secs': 1.467288} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.290547] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 866.290759] env[62965]: INFO nova.compute.manager [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Took 7.80 seconds to spawn the instance on the hypervisor. [ 866.290959] env[62965]: DEBUG nova.compute.manager [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 866.291761] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-278fb537-8cbc-4e84-bf0e-6f23acfcfc18 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.297013] env[62965]: DEBUG nova.policy [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '34e15ad3c87445e5a217300a23223540', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8bfa5cfdc7d5496482d87675f2a122dd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 866.544459] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Acquiring lock "refresh_cache-924c5cdc-0450-4ced-b920-50e2d5060fd2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.544727] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Acquired lock "refresh_cache-924c5cdc-0450-4ced-b920-50e2d5060fd2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.544808] env[62965]: DEBUG nova.network.neutron [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 866.570578] env[62965]: DEBUG oslo_vmware.api [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]527150bc-94d2-07b7-214c-826588c6dfbd, 'name': SearchDatastore_Task, 'duration_secs': 0.050896} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.570898] env[62965]: DEBUG oslo_concurrency.lockutils [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.571142] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 866.571370] env[62965]: DEBUG oslo_concurrency.lockutils [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.571516] env[62965]: DEBUG oslo_concurrency.lockutils [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.571713] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 866.573626] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-989c6a32-f1d0-40f4-9763-2cdbb118ed04 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.580120] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 866.580301] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 866.581040] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f6bf1b6-df55-4d20-87a8-ed2c3a66c1c9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.587015] env[62965]: DEBUG oslo_vmware.api [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 866.587015] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5298246f-5acb-260d-065e-be746c6c3c16" [ 866.587015] env[62965]: _type = "Task" [ 866.587015] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.594573] env[62965]: DEBUG oslo_vmware.api [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5298246f-5acb-260d-065e-be746c6c3c16, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.602530] env[62965]: DEBUG nova.network.neutron [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Successfully created port: c0f50257-3550-4368-8ad1-aa07da3d6e67 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 866.682513] env[62965]: DEBUG oslo_vmware.api [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389849, 'name': ReconfigVM_Task, 'duration_secs': 0.283389} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.685228] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Reconfigured VM instance instance-00000043 to attach disk [datastore2] d4d9b202-3aa7-4b66-949d-4a2eb34b3adf/d4d9b202-3aa7-4b66-949d-4a2eb34b3adf.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 866.686403] env[62965]: DEBUG oslo_vmware.api [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': task-1389850, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.176057} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.686642] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0272e58e-4036-4e50-86e1-d55dde482648 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.688268] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 866.688451] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Deleted contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 866.688627] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 866.695757] env[62965]: DEBUG oslo_vmware.api [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 866.695757] env[62965]: value = "task-1389851" [ 866.695757] env[62965]: _type = "Task" [ 866.695757] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.704481] env[62965]: DEBUG oslo_vmware.api [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389851, 'name': Rename_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.740288] env[62965]: DEBUG oslo_concurrency.lockutils [req-079d143f-b877-4aa5-8e41-6133f86e93e9 req-d8146b69-973b-4ae0-bc7e-7fb304b3299d service nova] Releasing lock "refresh_cache-b0d8f38a-86d3-466a-9b83-35145071dc21" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.756912] env[62965]: DEBUG nova.compute.manager [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 866.815013] env[62965]: INFO nova.compute.manager [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Took 27.65 seconds to build instance. [ 866.962571] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15b7672e-a899-43f9-a30c-59a40568bf7f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.969648] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c061917a-9b48-45c5-a559-42dae212d51f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.005343] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a463654c-eb4f-4d73-b5a6-5744fcd19afa {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.012216] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ddec0e6-8601-40eb-b765-fbbefd6e4836 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.027391] env[62965]: DEBUG nova.compute.provider_tree [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 867.097302] env[62965]: DEBUG oslo_vmware.api [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5298246f-5acb-260d-065e-be746c6c3c16, 'name': SearchDatastore_Task, 'duration_secs': 0.008367} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.097992] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e2f2def-f454-47ad-a433-4dc527026e50 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.103700] env[62965]: DEBUG oslo_vmware.api [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 867.103700] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52ead6f0-06e4-ba8f-1199-b64e9e0692c4" [ 867.103700] env[62965]: _type = "Task" [ 867.103700] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.111125] env[62965]: DEBUG oslo_vmware.api [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52ead6f0-06e4-ba8f-1199-b64e9e0692c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.117825] env[62965]: DEBUG nova.network.neutron [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 867.195962] env[62965]: DEBUG nova.compute.manager [req-96899975-1ce6-482d-bf3d-f097cf02dd37 req-97cd3686-9079-46fa-a182-b28d6200dd4c service nova] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Received event network-changed-57672e8b-1df1-463d-b9b3-8290667d9fd2 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 867.196158] env[62965]: DEBUG nova.compute.manager [req-96899975-1ce6-482d-bf3d-f097cf02dd37 req-97cd3686-9079-46fa-a182-b28d6200dd4c service nova] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Refreshing instance network info cache due to event network-changed-57672e8b-1df1-463d-b9b3-8290667d9fd2. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 867.196365] env[62965]: DEBUG oslo_concurrency.lockutils [req-96899975-1ce6-482d-bf3d-f097cf02dd37 req-97cd3686-9079-46fa-a182-b28d6200dd4c service nova] Acquiring lock "refresh_cache-c2448b76-1553-4b68-a731-0bb0967f4c1d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.196497] env[62965]: DEBUG oslo_concurrency.lockutils [req-96899975-1ce6-482d-bf3d-f097cf02dd37 req-97cd3686-9079-46fa-a182-b28d6200dd4c service nova] Acquired lock "refresh_cache-c2448b76-1553-4b68-a731-0bb0967f4c1d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.196658] env[62965]: DEBUG nova.network.neutron [req-96899975-1ce6-482d-bf3d-f097cf02dd37 req-97cd3686-9079-46fa-a182-b28d6200dd4c service nova] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Refreshing network info cache for port 57672e8b-1df1-463d-b9b3-8290667d9fd2 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 867.214355] env[62965]: DEBUG oslo_vmware.api [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389851, 'name': Rename_Task, 'duration_secs': 0.133689} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.216784] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 867.217042] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-88249d75-8137-42d1-83a3-fd8293d267b8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.223692] env[62965]: DEBUG oslo_vmware.api [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 867.223692] env[62965]: value = "task-1389852" [ 867.223692] env[62965]: _type = "Task" [ 867.223692] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.231088] env[62965]: DEBUG oslo_vmware.api [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389852, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.296162] env[62965]: DEBUG nova.compute.manager [req-3cf5b3a3-4168-4166-ae86-cdce96b1c30b req-6eac817b-b511-464f-bc60-5c2de93ec671 service nova] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Received event network-vif-plugged-80afc0d2-0d83-4374-a4c0-792eabf66c40 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 867.296162] env[62965]: DEBUG oslo_concurrency.lockutils [req-3cf5b3a3-4168-4166-ae86-cdce96b1c30b req-6eac817b-b511-464f-bc60-5c2de93ec671 service nova] Acquiring lock "924c5cdc-0450-4ced-b920-50e2d5060fd2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.296908] env[62965]: DEBUG oslo_concurrency.lockutils [req-3cf5b3a3-4168-4166-ae86-cdce96b1c30b req-6eac817b-b511-464f-bc60-5c2de93ec671 service nova] Lock "924c5cdc-0450-4ced-b920-50e2d5060fd2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.297496] env[62965]: DEBUG oslo_concurrency.lockutils [req-3cf5b3a3-4168-4166-ae86-cdce96b1c30b req-6eac817b-b511-464f-bc60-5c2de93ec671 service nova] Lock "924c5cdc-0450-4ced-b920-50e2d5060fd2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.298045] env[62965]: DEBUG nova.compute.manager [req-3cf5b3a3-4168-4166-ae86-cdce96b1c30b req-6eac817b-b511-464f-bc60-5c2de93ec671 service nova] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] No waiting events found dispatching network-vif-plugged-80afc0d2-0d83-4374-a4c0-792eabf66c40 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 867.299575] env[62965]: WARNING nova.compute.manager [req-3cf5b3a3-4168-4166-ae86-cdce96b1c30b req-6eac817b-b511-464f-bc60-5c2de93ec671 service nova] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Received unexpected event network-vif-plugged-80afc0d2-0d83-4374-a4c0-792eabf66c40 for instance with vm_state building and task_state spawning. [ 867.299575] env[62965]: DEBUG nova.compute.manager [req-3cf5b3a3-4168-4166-ae86-cdce96b1c30b req-6eac817b-b511-464f-bc60-5c2de93ec671 service nova] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Received event network-changed-80afc0d2-0d83-4374-a4c0-792eabf66c40 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 867.299575] env[62965]: DEBUG nova.compute.manager [req-3cf5b3a3-4168-4166-ae86-cdce96b1c30b req-6eac817b-b511-464f-bc60-5c2de93ec671 service nova] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Refreshing instance network info cache due to event network-changed-80afc0d2-0d83-4374-a4c0-792eabf66c40. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 867.299575] env[62965]: DEBUG oslo_concurrency.lockutils [req-3cf5b3a3-4168-4166-ae86-cdce96b1c30b req-6eac817b-b511-464f-bc60-5c2de93ec671 service nova] Acquiring lock "refresh_cache-924c5cdc-0450-4ced-b920-50e2d5060fd2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.321035] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1d916132-970a-4551-af56-4b52e2162d14 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "c2448b76-1553-4b68-a731-0bb0967f4c1d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 82.204s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.343941] env[62965]: DEBUG nova.network.neutron [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Updating instance_info_cache with network_info: [{"id": "80afc0d2-0d83-4374-a4c0-792eabf66c40", "address": "fa:16:3e:1c:0d:5e", "network": {"id": "7ef06255-01ec-4620-8f3d-6085875aa8a3", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-118464573-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1caebff08644e21b3f1c461aa115b8e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap80afc0d2-0d", "ovs_interfaceid": "80afc0d2-0d83-4374-a4c0-792eabf66c40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.530014] env[62965]: DEBUG nova.scheduler.client.report [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 867.614480] env[62965]: DEBUG oslo_vmware.api [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52ead6f0-06e4-ba8f-1199-b64e9e0692c4, 'name': SearchDatastore_Task, 'duration_secs': 0.008989} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.614761] env[62965]: DEBUG oslo_concurrency.lockutils [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.615030] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] b0d8f38a-86d3-466a-9b83-35145071dc21/b0d8f38a-86d3-466a-9b83-35145071dc21.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 867.615290] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5e90e634-4a5b-499c-b595-3d47dfeff417 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.621974] env[62965]: DEBUG oslo_vmware.api [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 867.621974] env[62965]: value = "task-1389853" [ 867.621974] env[62965]: _type = "Task" [ 867.621974] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.629849] env[62965]: DEBUG oslo_vmware.api [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1389853, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.727054] env[62965]: DEBUG nova.virt.hardware [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 867.727274] env[62965]: DEBUG nova.virt.hardware [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 867.727442] env[62965]: DEBUG nova.virt.hardware [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 867.727703] env[62965]: DEBUG nova.virt.hardware [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 867.727941] env[62965]: DEBUG nova.virt.hardware [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 867.728019] env[62965]: DEBUG nova.virt.hardware [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 867.728180] env[62965]: DEBUG nova.virt.hardware [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 867.728337] env[62965]: DEBUG nova.virt.hardware [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 867.728501] env[62965]: DEBUG nova.virt.hardware [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 867.728658] env[62965]: DEBUG nova.virt.hardware [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 867.728832] env[62965]: DEBUG nova.virt.hardware [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 867.729950] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0c35523-f1b7-4c53-b8fe-1a2ad53d491d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.745073] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09b5403b-7e3e-45e1-a506-7c32d466d1e0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.749024] env[62965]: DEBUG oslo_vmware.api [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389852, 'name': PowerOnVM_Task, 'duration_secs': 0.446458} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.749318] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 867.749567] env[62965]: INFO nova.compute.manager [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Took 6.97 seconds to spawn the instance on the hypervisor. [ 867.749760] env[62965]: DEBUG nova.compute.manager [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 867.751247] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-769dc23e-8e98-497f-a388-61a6dccdda9c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.761590] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Instance VIF info [] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 867.767064] env[62965]: DEBUG oslo.service.loopingcall [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 867.768547] env[62965]: DEBUG nova.compute.manager [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 867.770539] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 867.772419] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e33c3be9-03a7-4287-8040-60fe2318b2ea {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.797123] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 867.797123] env[62965]: value = "task-1389854" [ 867.797123] env[62965]: _type = "Task" [ 867.797123] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.799263] env[62965]: DEBUG nova.virt.hardware [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 867.799489] env[62965]: DEBUG nova.virt.hardware [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 867.799664] env[62965]: DEBUG nova.virt.hardware [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 867.799857] env[62965]: DEBUG nova.virt.hardware [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 867.799999] env[62965]: DEBUG nova.virt.hardware [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 867.800161] env[62965]: DEBUG nova.virt.hardware [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 867.800364] env[62965]: DEBUG nova.virt.hardware [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 867.800515] env[62965]: DEBUG nova.virt.hardware [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 867.800680] env[62965]: DEBUG nova.virt.hardware [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 867.800841] env[62965]: DEBUG nova.virt.hardware [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 867.801048] env[62965]: DEBUG nova.virt.hardware [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 867.801896] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-746191ca-6321-464e-8eb0-fb8506466f1c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.814118] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389854, 'name': CreateVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.817514] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb49021a-d43e-4026-b7ea-17cd8c5cf802 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.823481] env[62965]: DEBUG nova.compute.manager [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 867.849389] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Releasing lock "refresh_cache-924c5cdc-0450-4ced-b920-50e2d5060fd2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.849792] env[62965]: DEBUG nova.compute.manager [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Instance network_info: |[{"id": "80afc0d2-0d83-4374-a4c0-792eabf66c40", "address": "fa:16:3e:1c:0d:5e", "network": {"id": "7ef06255-01ec-4620-8f3d-6085875aa8a3", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-118464573-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1caebff08644e21b3f1c461aa115b8e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap80afc0d2-0d", "ovs_interfaceid": "80afc0d2-0d83-4374-a4c0-792eabf66c40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 867.850133] env[62965]: DEBUG oslo_concurrency.lockutils [req-3cf5b3a3-4168-4166-ae86-cdce96b1c30b req-6eac817b-b511-464f-bc60-5c2de93ec671 service nova] Acquired lock "refresh_cache-924c5cdc-0450-4ced-b920-50e2d5060fd2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.850448] env[62965]: DEBUG nova.network.neutron [req-3cf5b3a3-4168-4166-ae86-cdce96b1c30b req-6eac817b-b511-464f-bc60-5c2de93ec671 service nova] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Refreshing network info cache for port 80afc0d2-0d83-4374-a4c0-792eabf66c40 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 867.851748] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1c:0d:5e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '03ac2c9c-6ad2-4a85-bfab-c7e336df859a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '80afc0d2-0d83-4374-a4c0-792eabf66c40', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 867.859726] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Creating folder: Project (c1caebff08644e21b3f1c461aa115b8e). Parent ref: group-v294931. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 867.861359] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7900520b-c836-4489-9c4a-9408f9bc1c88 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.875418] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Created folder: Project (c1caebff08644e21b3f1c461aa115b8e) in parent group-v294931. [ 867.875642] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Creating folder: Instances. Parent ref: group-v294969. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 867.875894] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-26024d5b-1d40-4b23-bf27-96fb6a5ee150 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.888404] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Created folder: Instances in parent group-v294969. [ 867.888878] env[62965]: DEBUG oslo.service.loopingcall [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 867.889008] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 867.889250] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-23ba1201-f348-4c50-866d-661ee40150ec {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.912428] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 867.912428] env[62965]: value = "task-1389857" [ 867.912428] env[62965]: _type = "Task" [ 867.912428] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.920275] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389857, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.020358] env[62965]: DEBUG nova.network.neutron [req-96899975-1ce6-482d-bf3d-f097cf02dd37 req-97cd3686-9079-46fa-a182-b28d6200dd4c service nova] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Updated VIF entry in instance network info cache for port 57672e8b-1df1-463d-b9b3-8290667d9fd2. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 868.021405] env[62965]: DEBUG nova.network.neutron [req-96899975-1ce6-482d-bf3d-f097cf02dd37 req-97cd3686-9079-46fa-a182-b28d6200dd4c service nova] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Updating instance_info_cache with network_info: [{"id": "57672e8b-1df1-463d-b9b3-8290667d9fd2", "address": "fa:16:3e:53:46:47", "network": {"id": "4649fed7-c2ca-43a9-a183-7c99423da726", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1978698173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fab418e2fe1420793517663574b43bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43ad01d2-c7dd-453c-a929-8ad76294d13c", "external-id": "nsx-vlan-transportzone-176", "segmentation_id": 176, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57672e8b-1d", "ovs_interfaceid": "57672e8b-1df1-463d-b9b3-8290667d9fd2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.036672] env[62965]: DEBUG oslo_concurrency.lockutils [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.292s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.037224] env[62965]: DEBUG nova.compute.manager [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 868.040118] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.646s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.133798] env[62965]: DEBUG oslo_vmware.api [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1389853, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.300911] env[62965]: INFO nova.compute.manager [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Took 27.30 seconds to build instance. [ 868.311823] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389854, 'name': CreateVM_Task} progress is 99%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.363019] env[62965]: DEBUG oslo_concurrency.lockutils [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.379507] env[62965]: DEBUG nova.network.neutron [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Successfully updated port: c0f50257-3550-4368-8ad1-aa07da3d6e67 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 868.421956] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389857, 'name': CreateVM_Task} progress is 99%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.523544] env[62965]: DEBUG oslo_concurrency.lockutils [req-96899975-1ce6-482d-bf3d-f097cf02dd37 req-97cd3686-9079-46fa-a182-b28d6200dd4c service nova] Releasing lock "refresh_cache-c2448b76-1553-4b68-a731-0bb0967f4c1d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.545364] env[62965]: DEBUG nova.compute.utils [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 868.551956] env[62965]: DEBUG nova.compute.manager [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 868.552038] env[62965]: DEBUG nova.network.neutron [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 868.595487] env[62965]: DEBUG nova.policy [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a5e5c6c3198946658126b0024ba82583', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9a539017d4b84f39a71b6f4098395fbe', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 868.630862] env[62965]: DEBUG nova.network.neutron [req-3cf5b3a3-4168-4166-ae86-cdce96b1c30b req-6eac817b-b511-464f-bc60-5c2de93ec671 service nova] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Updated VIF entry in instance network info cache for port 80afc0d2-0d83-4374-a4c0-792eabf66c40. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 868.631242] env[62965]: DEBUG nova.network.neutron [req-3cf5b3a3-4168-4166-ae86-cdce96b1c30b req-6eac817b-b511-464f-bc60-5c2de93ec671 service nova] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Updating instance_info_cache with network_info: [{"id": "80afc0d2-0d83-4374-a4c0-792eabf66c40", "address": "fa:16:3e:1c:0d:5e", "network": {"id": "7ef06255-01ec-4620-8f3d-6085875aa8a3", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-118464573-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1caebff08644e21b3f1c461aa115b8e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap80afc0d2-0d", "ovs_interfaceid": "80afc0d2-0d83-4374-a4c0-792eabf66c40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.638394] env[62965]: DEBUG oslo_vmware.api [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1389853, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.525785} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.642086] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] b0d8f38a-86d3-466a-9b83-35145071dc21/b0d8f38a-86d3-466a-9b83-35145071dc21.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 868.642086] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 868.642086] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2384ae4f-3443-4733-82e7-e1f78e4c11c4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.649086] env[62965]: DEBUG oslo_vmware.api [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 868.649086] env[62965]: value = "task-1389858" [ 868.649086] env[62965]: _type = "Task" [ 868.649086] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.661141] env[62965]: DEBUG oslo_vmware.api [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1389858, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.786424] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e63b73bc-d8a4-4a93-b9d5-da54dffe81b4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.800790] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65b96110-9588-4530-a497-558c2dc6942e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.808761] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ed6b41cd-900a-47fd-91cb-8865fabec32f tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "d4d9b202-3aa7-4b66-949d-4a2eb34b3adf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.157s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.844669] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389854, 'name': CreateVM_Task, 'duration_secs': 0.536444} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.845738] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 868.846503] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a7b3645-7e40-4b71-a003-95f7bd81f0f4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.850072] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.850072] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.850072] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 868.850423] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-945f3e6c-9ea6-4e8e-8d4c-2a5f81f30f44 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.857827] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-246be253-f4fa-467b-890c-b9483db42de5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.861790] env[62965]: DEBUG oslo_vmware.api [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Waiting for the task: (returnval){ [ 868.861790] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5211e264-6a9d-811b-c438-210a4c7097cd" [ 868.861790] env[62965]: _type = "Task" [ 868.861790] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.875232] env[62965]: DEBUG nova.compute.provider_tree [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 868.880376] env[62965]: DEBUG oslo_vmware.api [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5211e264-6a9d-811b-c438-210a4c7097cd, 'name': SearchDatastore_Task, 'duration_secs': 0.012149} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.880673] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.880899] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 868.881142] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.881286] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.881500] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 868.881943] env[62965]: DEBUG oslo_concurrency.lockutils [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "refresh_cache-8c478341-9b50-45de-9ba3-321e802ed1f0" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.882078] env[62965]: DEBUG oslo_concurrency.lockutils [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquired lock "refresh_cache-8c478341-9b50-45de-9ba3-321e802ed1f0" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.882203] env[62965]: DEBUG nova.network.neutron [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 868.883629] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7a2662b1-0098-4659-9ad5-c5c5a78b7124 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.892754] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 868.892754] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 868.892754] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d7d5170-65d6-4085-a83e-ee16e1317e25 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.898881] env[62965]: DEBUG oslo_vmware.api [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Waiting for the task: (returnval){ [ 868.898881] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5240cfa0-d137-9dd1-c757-73081def27f1" [ 868.898881] env[62965]: _type = "Task" [ 868.898881] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.902944] env[62965]: DEBUG nova.network.neutron [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Successfully created port: 65a91c43-8ff0-4e66-8f18-2de456875968 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 868.913019] env[62965]: DEBUG oslo_vmware.api [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5240cfa0-d137-9dd1-c757-73081def27f1, 'name': SearchDatastore_Task, 'duration_secs': 0.008757} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.913502] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69bc453d-7798-4879-9c03-d87ad3fd8d58 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.922368] env[62965]: DEBUG oslo_vmware.api [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Waiting for the task: (returnval){ [ 868.922368] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52e4ba9e-234b-29d6-0def-ba473c718d1e" [ 868.922368] env[62965]: _type = "Task" [ 868.922368] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.925967] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389857, 'name': CreateVM_Task, 'duration_secs': 0.581952} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.929187] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 868.929879] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.930053] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.930364] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 868.931197] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d3c6af39-797d-4019-80e4-f392ba2ccd48 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.936261] env[62965]: DEBUG oslo_vmware.api [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52e4ba9e-234b-29d6-0def-ba473c718d1e, 'name': SearchDatastore_Task, 'duration_secs': 0.008458} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.936817] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.937084] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c/d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 868.937363] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4893e1b9-c507-4a3b-88c9-ad0aa3b29ff6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.942686] env[62965]: DEBUG oslo_vmware.api [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Waiting for the task: (returnval){ [ 868.942686] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52ba9672-5d70-6e77-f81f-da93aaa7784c" [ 868.942686] env[62965]: _type = "Task" [ 868.942686] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.947409] env[62965]: DEBUG oslo_vmware.api [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Waiting for the task: (returnval){ [ 868.947409] env[62965]: value = "task-1389859" [ 868.947409] env[62965]: _type = "Task" [ 868.947409] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.955632] env[62965]: DEBUG oslo_vmware.api [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52ba9672-5d70-6e77-f81f-da93aaa7784c, 'name': SearchDatastore_Task, 'duration_secs': 0.009965} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.956330] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.960088] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 868.960393] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.960552] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.964017] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 868.964289] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b0a5bd12-4d38-4fc4-a172-208142fc1c4c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.966166] env[62965]: DEBUG oslo_vmware.api [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': task-1389859, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.971736] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 868.971929] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 868.972658] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e8f7785-459f-494d-b42d-7be9efcdc962 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.977978] env[62965]: DEBUG oslo_vmware.api [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Waiting for the task: (returnval){ [ 868.977978] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52529c80-e8a6-0c81-228e-c19d4e187e47" [ 868.977978] env[62965]: _type = "Task" [ 868.977978] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.985963] env[62965]: DEBUG oslo_vmware.api [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52529c80-e8a6-0c81-228e-c19d4e187e47, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.052666] env[62965]: DEBUG nova.compute.manager [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 869.134031] env[62965]: DEBUG oslo_concurrency.lockutils [req-3cf5b3a3-4168-4166-ae86-cdce96b1c30b req-6eac817b-b511-464f-bc60-5c2de93ec671 service nova] Releasing lock "refresh_cache-924c5cdc-0450-4ced-b920-50e2d5060fd2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.159351] env[62965]: DEBUG oslo_vmware.api [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1389858, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.146041} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.159776] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 869.160543] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c87a174-a494-4bb3-9b52-6e6b90c77281 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.189076] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] b0d8f38a-86d3-466a-9b83-35145071dc21/b0d8f38a-86d3-466a-9b83-35145071dc21.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 869.189943] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ae74b9d0-03e0-40f4-bc22-487c4772004d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.211658] env[62965]: DEBUG oslo_vmware.api [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 869.211658] env[62965]: value = "task-1389860" [ 869.211658] env[62965]: _type = "Task" [ 869.211658] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.224118] env[62965]: DEBUG oslo_vmware.api [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1389860, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.228324] env[62965]: DEBUG nova.compute.manager [req-28092d41-da55-45d9-9272-2475cc15e9eb req-e24a2c9d-2c47-4c0b-bb3e-67d5933eb1c7 service nova] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Received event network-vif-plugged-c0f50257-3550-4368-8ad1-aa07da3d6e67 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 869.228498] env[62965]: DEBUG oslo_concurrency.lockutils [req-28092d41-da55-45d9-9272-2475cc15e9eb req-e24a2c9d-2c47-4c0b-bb3e-67d5933eb1c7 service nova] Acquiring lock "8c478341-9b50-45de-9ba3-321e802ed1f0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.228652] env[62965]: DEBUG oslo_concurrency.lockutils [req-28092d41-da55-45d9-9272-2475cc15e9eb req-e24a2c9d-2c47-4c0b-bb3e-67d5933eb1c7 service nova] Lock "8c478341-9b50-45de-9ba3-321e802ed1f0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.228809] env[62965]: DEBUG oslo_concurrency.lockutils [req-28092d41-da55-45d9-9272-2475cc15e9eb req-e24a2c9d-2c47-4c0b-bb3e-67d5933eb1c7 service nova] Lock "8c478341-9b50-45de-9ba3-321e802ed1f0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.228970] env[62965]: DEBUG nova.compute.manager [req-28092d41-da55-45d9-9272-2475cc15e9eb req-e24a2c9d-2c47-4c0b-bb3e-67d5933eb1c7 service nova] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] No waiting events found dispatching network-vif-plugged-c0f50257-3550-4368-8ad1-aa07da3d6e67 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 869.229244] env[62965]: WARNING nova.compute.manager [req-28092d41-da55-45d9-9272-2475cc15e9eb req-e24a2c9d-2c47-4c0b-bb3e-67d5933eb1c7 service nova] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Received unexpected event network-vif-plugged-c0f50257-3550-4368-8ad1-aa07da3d6e67 for instance with vm_state building and task_state spawning. [ 869.229443] env[62965]: DEBUG nova.compute.manager [req-28092d41-da55-45d9-9272-2475cc15e9eb req-e24a2c9d-2c47-4c0b-bb3e-67d5933eb1c7 service nova] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Received event network-changed-c0f50257-3550-4368-8ad1-aa07da3d6e67 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 869.229797] env[62965]: DEBUG nova.compute.manager [req-28092d41-da55-45d9-9272-2475cc15e9eb req-e24a2c9d-2c47-4c0b-bb3e-67d5933eb1c7 service nova] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Refreshing instance network info cache due to event network-changed-c0f50257-3550-4368-8ad1-aa07da3d6e67. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 869.230017] env[62965]: DEBUG oslo_concurrency.lockutils [req-28092d41-da55-45d9-9272-2475cc15e9eb req-e24a2c9d-2c47-4c0b-bb3e-67d5933eb1c7 service nova] Acquiring lock "refresh_cache-8c478341-9b50-45de-9ba3-321e802ed1f0" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.313394] env[62965]: DEBUG nova.compute.manager [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 869.377443] env[62965]: DEBUG nova.scheduler.client.report [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 869.430661] env[62965]: DEBUG nova.network.neutron [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 869.459152] env[62965]: DEBUG oslo_vmware.api [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': task-1389859, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.490100] env[62965]: DEBUG oslo_vmware.api [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52529c80-e8a6-0c81-228e-c19d4e187e47, 'name': SearchDatastore_Task, 'duration_secs': 0.008411} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.490715] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c624ece5-3512-4910-8301-5c8025c355ef {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.495690] env[62965]: DEBUG oslo_vmware.api [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Waiting for the task: (returnval){ [ 869.495690] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52022e6e-4525-4874-7abe-445e0eaa6653" [ 869.495690] env[62965]: _type = "Task" [ 869.495690] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.505087] env[62965]: DEBUG oslo_vmware.api [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52022e6e-4525-4874-7abe-445e0eaa6653, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.600229] env[62965]: DEBUG nova.network.neutron [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Updating instance_info_cache with network_info: [{"id": "c0f50257-3550-4368-8ad1-aa07da3d6e67", "address": "fa:16:3e:0d:28:72", "network": {"id": "a5246cc1-6001-4c20-9724-fd36a0437582", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-644702631-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bfa5cfdc7d5496482d87675f2a122dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0f50257-35", "ovs_interfaceid": "c0f50257-3550-4368-8ad1-aa07da3d6e67", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.722067] env[62965]: DEBUG oslo_vmware.api [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1389860, 'name': ReconfigVM_Task, 'duration_secs': 0.463491} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.722067] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Reconfigured VM instance instance-00000044 to attach disk [datastore2] b0d8f38a-86d3-466a-9b83-35145071dc21/b0d8f38a-86d3-466a-9b83-35145071dc21.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 869.722530] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fee7958f-5f28-438b-a157-a2ee30be8915 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.727955] env[62965]: DEBUG oslo_vmware.api [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 869.727955] env[62965]: value = "task-1389861" [ 869.727955] env[62965]: _type = "Task" [ 869.727955] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.735061] env[62965]: DEBUG oslo_vmware.api [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1389861, 'name': Rename_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.838287] env[62965]: DEBUG oslo_concurrency.lockutils [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.883845] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.844s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.884517] env[62965]: ERROR nova.compute.manager [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2c3621b8-53b9-48d0-90ca-6f9450d91d7a, please check neutron logs for more information. [ 869.884517] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Traceback (most recent call last): [ 869.884517] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 869.884517] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] self.driver.spawn(context, instance, image_meta, [ 869.884517] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 548, in spawn [ 869.884517] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 869.884517] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 869.884517] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] vm_ref = self.build_virtual_machine(instance, [ 869.884517] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 869.884517] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] vif_infos = vmwarevif.get_vif_info(self._session, [ 869.884517] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 869.884914] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] for vif in network_info: [ 869.884914] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 869.884914] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] return self._sync_wrapper(fn, *args, **kwargs) [ 869.884914] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 869.884914] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] self.wait() [ 869.884914] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 869.884914] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] self[:] = self._gt.wait() [ 869.884914] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 869.884914] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] return self._exit_event.wait() [ 869.884914] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 869.884914] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] result = hub.switch() [ 869.884914] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 869.884914] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] return self.greenlet.switch() [ 869.885290] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 869.885290] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] result = function(*args, **kwargs) [ 869.885290] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 869.885290] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] return func(*args, **kwargs) [ 869.885290] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 869.885290] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] raise e [ 869.885290] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 869.885290] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] nwinfo = self.network_api.allocate_for_instance( [ 869.885290] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 869.885290] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] created_port_ids = self._update_ports_for_instance( [ 869.885290] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 869.885290] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] with excutils.save_and_reraise_exception(): [ 869.885290] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 869.885643] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] self.force_reraise() [ 869.885643] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 869.885643] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] raise self.value [ 869.885643] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 869.885643] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] updated_port = self._update_port( [ 869.885643] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 869.885643] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] _ensure_no_port_binding_failure(port) [ 869.885643] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 869.885643] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] raise exception.PortBindingFailed(port_id=port['id']) [ 869.885643] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] nova.exception.PortBindingFailed: Binding failed for port 2c3621b8-53b9-48d0-90ca-6f9450d91d7a, please check neutron logs for more information. [ 869.885643] env[62965]: ERROR nova.compute.manager [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] [ 869.885956] env[62965]: DEBUG nova.compute.utils [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Binding failed for port 2c3621b8-53b9-48d0-90ca-6f9450d91d7a, please check neutron logs for more information. {{(pid=62965) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 869.886937] env[62965]: DEBUG oslo_concurrency.lockutils [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.277s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.888750] env[62965]: INFO nova.compute.claims [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 869.891487] env[62965]: DEBUG nova.compute.manager [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Build of instance 2b6efa32-fc5b-402c-a551-b6e5c01eca2c was re-scheduled: Binding failed for port 2c3621b8-53b9-48d0-90ca-6f9450d91d7a, please check neutron logs for more information. {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 869.891912] env[62965]: DEBUG nova.compute.manager [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Unplugging VIFs for instance {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 869.892162] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] Acquiring lock "refresh_cache-2b6efa32-fc5b-402c-a551-b6e5c01eca2c" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.892307] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] Acquired lock "refresh_cache-2b6efa32-fc5b-402c-a551-b6e5c01eca2c" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.892464] env[62965]: DEBUG nova.network.neutron [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 869.918465] env[62965]: DEBUG nova.compute.manager [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 869.919965] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64695b01-79c6-42e4-a19c-136e6adfdd78 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.959885] env[62965]: DEBUG oslo_vmware.api [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': task-1389859, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.514728} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.960147] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c/d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 869.960391] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 869.960922] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a6406f69-e32c-42ca-97ea-621b3bf461a0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.967552] env[62965]: DEBUG oslo_vmware.api [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Waiting for the task: (returnval){ [ 869.967552] env[62965]: value = "task-1389862" [ 869.967552] env[62965]: _type = "Task" [ 869.967552] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.975362] env[62965]: DEBUG oslo_vmware.api [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': task-1389862, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.005901] env[62965]: DEBUG oslo_vmware.api [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52022e6e-4525-4874-7abe-445e0eaa6653, 'name': SearchDatastore_Task, 'duration_secs': 0.009735} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.006181] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.006433] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 924c5cdc-0450-4ced-b920-50e2d5060fd2/924c5cdc-0450-4ced-b920-50e2d5060fd2.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 870.006692] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c154e567-b120-4770-a92e-6695b64c104f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.013573] env[62965]: DEBUG oslo_vmware.api [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Waiting for the task: (returnval){ [ 870.013573] env[62965]: value = "task-1389863" [ 870.013573] env[62965]: _type = "Task" [ 870.013573] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.021764] env[62965]: DEBUG oslo_vmware.api [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Task: {'id': task-1389863, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.065141] env[62965]: DEBUG nova.compute.manager [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 870.087456] env[62965]: DEBUG nova.virt.hardware [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 870.087694] env[62965]: DEBUG nova.virt.hardware [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 870.087850] env[62965]: DEBUG nova.virt.hardware [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 870.088046] env[62965]: DEBUG nova.virt.hardware [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 870.088198] env[62965]: DEBUG nova.virt.hardware [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 870.088342] env[62965]: DEBUG nova.virt.hardware [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 870.088547] env[62965]: DEBUG nova.virt.hardware [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 870.088707] env[62965]: DEBUG nova.virt.hardware [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 870.088874] env[62965]: DEBUG nova.virt.hardware [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 870.089053] env[62965]: DEBUG nova.virt.hardware [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 870.089239] env[62965]: DEBUG nova.virt.hardware [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 870.090438] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bd97c75-b12a-44f9-8dc8-03ef807fbdb6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.098393] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79b4e2f8-9661-473b-a60d-79c291215bfd {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.102323] env[62965]: DEBUG oslo_concurrency.lockutils [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Releasing lock "refresh_cache-8c478341-9b50-45de-9ba3-321e802ed1f0" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 870.102613] env[62965]: DEBUG nova.compute.manager [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Instance network_info: |[{"id": "c0f50257-3550-4368-8ad1-aa07da3d6e67", "address": "fa:16:3e:0d:28:72", "network": {"id": "a5246cc1-6001-4c20-9724-fd36a0437582", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-644702631-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bfa5cfdc7d5496482d87675f2a122dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0f50257-35", "ovs_interfaceid": "c0f50257-3550-4368-8ad1-aa07da3d6e67", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 870.102889] env[62965]: DEBUG oslo_concurrency.lockutils [req-28092d41-da55-45d9-9272-2475cc15e9eb req-e24a2c9d-2c47-4c0b-bb3e-67d5933eb1c7 service nova] Acquired lock "refresh_cache-8c478341-9b50-45de-9ba3-321e802ed1f0" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.103074] env[62965]: DEBUG nova.network.neutron [req-28092d41-da55-45d9-9272-2475cc15e9eb req-e24a2c9d-2c47-4c0b-bb3e-67d5933eb1c7 service nova] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Refreshing network info cache for port c0f50257-3550-4368-8ad1-aa07da3d6e67 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 870.104208] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0d:28:72', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '61b8f0db-488e-42d7-bf6c-6c1665cd5616', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c0f50257-3550-4368-8ad1-aa07da3d6e67', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 870.112460] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Creating folder: Project (8bfa5cfdc7d5496482d87675f2a122dd). Parent ref: group-v294931. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 870.114032] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-120f94a7-8f9e-49d0-b816-df2e7cc573a9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.132642] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Created folder: Project (8bfa5cfdc7d5496482d87675f2a122dd) in parent group-v294931. [ 870.132824] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Creating folder: Instances. Parent ref: group-v294972. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 870.133071] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9b7dfddc-c911-4d45-a5ac-5dd94179d098 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.143168] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Created folder: Instances in parent group-v294972. [ 870.143426] env[62965]: DEBUG oslo.service.loopingcall [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 870.143634] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 870.143859] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b7f95110-615f-4730-9953-dbec93bd5836 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.165919] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 870.165919] env[62965]: value = "task-1389866" [ 870.165919] env[62965]: _type = "Task" [ 870.165919] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.176098] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389866, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.238058] env[62965]: DEBUG oslo_vmware.api [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1389861, 'name': Rename_Task, 'duration_secs': 0.323767} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.238058] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 870.238329] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3b489e87-ab99-4a4c-9fe1-f508cfe4623d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.247341] env[62965]: DEBUG oslo_vmware.api [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 870.247341] env[62965]: value = "task-1389867" [ 870.247341] env[62965]: _type = "Task" [ 870.247341] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.257470] env[62965]: DEBUG oslo_vmware.api [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1389867, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.413095] env[62965]: DEBUG nova.network.neutron [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 870.431666] env[62965]: INFO nova.compute.manager [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] instance snapshotting [ 870.434849] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5506d08-2a44-48bb-8deb-16206a56f1f5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.455787] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc21c5c3-7857-4561-a4e9-b2439d88006d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.482202] env[62965]: DEBUG oslo_vmware.api [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': task-1389862, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.176408} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.482619] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 870.483773] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-476c9a32-c408-4c42-b471-9c095f829e96 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.514760] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c/d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 870.516194] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0ebdb8a7-b040-4dc2-ab82-b62c1d1f42f2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.541377] env[62965]: DEBUG oslo_vmware.api [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Task: {'id': task-1389863, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.543109] env[62965]: DEBUG oslo_vmware.api [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Waiting for the task: (returnval){ [ 870.543109] env[62965]: value = "task-1389868" [ 870.543109] env[62965]: _type = "Task" [ 870.543109] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.552452] env[62965]: DEBUG oslo_vmware.api [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': task-1389868, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.585053] env[62965]: DEBUG nova.network.neutron [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Successfully updated port: 65a91c43-8ff0-4e66-8f18-2de456875968 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 870.594149] env[62965]: DEBUG nova.network.neutron [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.675930] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389866, 'name': CreateVM_Task} progress is 25%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.761836] env[62965]: DEBUG oslo_vmware.api [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1389867, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.912383] env[62965]: DEBUG nova.network.neutron [req-28092d41-da55-45d9-9272-2475cc15e9eb req-e24a2c9d-2c47-4c0b-bb3e-67d5933eb1c7 service nova] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Updated VIF entry in instance network info cache for port c0f50257-3550-4368-8ad1-aa07da3d6e67. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 870.912785] env[62965]: DEBUG nova.network.neutron [req-28092d41-da55-45d9-9272-2475cc15e9eb req-e24a2c9d-2c47-4c0b-bb3e-67d5933eb1c7 service nova] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Updating instance_info_cache with network_info: [{"id": "c0f50257-3550-4368-8ad1-aa07da3d6e67", "address": "fa:16:3e:0d:28:72", "network": {"id": "a5246cc1-6001-4c20-9724-fd36a0437582", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-644702631-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bfa5cfdc7d5496482d87675f2a122dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0f50257-35", "ovs_interfaceid": "c0f50257-3550-4368-8ad1-aa07da3d6e67", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.971796] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Creating Snapshot of the VM instance {{(pid=62965) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 870.971893] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-31205435-e3fd-458c-b987-91e7eb29bd72 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.979216] env[62965]: DEBUG oslo_vmware.api [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 870.979216] env[62965]: value = "task-1389869" [ 870.979216] env[62965]: _type = "Task" [ 870.979216] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.987119] env[62965]: DEBUG oslo_vmware.api [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389869, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.027937] env[62965]: DEBUG oslo_vmware.api [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Task: {'id': task-1389863, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.530165} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.028325] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 924c5cdc-0450-4ced-b920-50e2d5060fd2/924c5cdc-0450-4ced-b920-50e2d5060fd2.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 871.028576] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 871.028834] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8577f36b-0f40-4f5d-b3c4-98343636d157 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.036365] env[62965]: DEBUG oslo_vmware.api [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Waiting for the task: (returnval){ [ 871.036365] env[62965]: value = "task-1389870" [ 871.036365] env[62965]: _type = "Task" [ 871.036365] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.044828] env[62965]: DEBUG oslo_vmware.api [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Task: {'id': task-1389870, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.056074] env[62965]: DEBUG oslo_vmware.api [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': task-1389868, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.085776] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21554039-7508-49a7-8ac7-a3569135e89b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.090493] env[62965]: DEBUG oslo_concurrency.lockutils [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquiring lock "refresh_cache-6def4a38-d3b8-4724-97a1-1859b239d900" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.090662] env[62965]: DEBUG oslo_concurrency.lockutils [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquired lock "refresh_cache-6def4a38-d3b8-4724-97a1-1859b239d900" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.090776] env[62965]: DEBUG nova.network.neutron [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 871.095233] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7113bb85-6f86-45ab-9552-8390d60ffc13 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.099167] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] Releasing lock "refresh_cache-2b6efa32-fc5b-402c-a551-b6e5c01eca2c" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.099398] env[62965]: DEBUG nova.compute.manager [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62965) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 871.099627] env[62965]: DEBUG nova.compute.manager [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 871.099818] env[62965]: DEBUG nova.network.neutron [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 871.131407] env[62965]: DEBUG nova.network.neutron [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 871.133374] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16c4b48c-7926-4c29-80db-96841a38f443 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.142426] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a98d9596-1e58-4428-8d35-97e128c07f42 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.157981] env[62965]: DEBUG nova.compute.provider_tree [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 871.175356] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389866, 'name': CreateVM_Task, 'duration_secs': 0.626756} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.175516] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 871.176204] env[62965]: DEBUG oslo_concurrency.lockutils [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.176370] env[62965]: DEBUG oslo_concurrency.lockutils [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.176672] env[62965]: DEBUG oslo_concurrency.lockutils [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 871.176918] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-94dcb845-9d45-46d4-9d95-46a76cd7f76b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.181395] env[62965]: DEBUG oslo_vmware.api [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 871.181395] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]525a7f33-b1ac-30fe-a397-7459aec8e675" [ 871.181395] env[62965]: _type = "Task" [ 871.181395] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.189092] env[62965]: DEBUG oslo_vmware.api [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]525a7f33-b1ac-30fe-a397-7459aec8e675, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.255518] env[62965]: DEBUG nova.compute.manager [req-c43cbdd5-9f03-4553-94e9-466d9f7cc004 req-f6797485-158c-4127-8481-ec03f6bfe4c5 service nova] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Received event network-vif-plugged-65a91c43-8ff0-4e66-8f18-2de456875968 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 871.255749] env[62965]: DEBUG oslo_concurrency.lockutils [req-c43cbdd5-9f03-4553-94e9-466d9f7cc004 req-f6797485-158c-4127-8481-ec03f6bfe4c5 service nova] Acquiring lock "6def4a38-d3b8-4724-97a1-1859b239d900-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.255997] env[62965]: DEBUG oslo_concurrency.lockutils [req-c43cbdd5-9f03-4553-94e9-466d9f7cc004 req-f6797485-158c-4127-8481-ec03f6bfe4c5 service nova] Lock "6def4a38-d3b8-4724-97a1-1859b239d900-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.256144] env[62965]: DEBUG oslo_concurrency.lockutils [req-c43cbdd5-9f03-4553-94e9-466d9f7cc004 req-f6797485-158c-4127-8481-ec03f6bfe4c5 service nova] Lock "6def4a38-d3b8-4724-97a1-1859b239d900-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.256308] env[62965]: DEBUG nova.compute.manager [req-c43cbdd5-9f03-4553-94e9-466d9f7cc004 req-f6797485-158c-4127-8481-ec03f6bfe4c5 service nova] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] No waiting events found dispatching network-vif-plugged-65a91c43-8ff0-4e66-8f18-2de456875968 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 871.256466] env[62965]: WARNING nova.compute.manager [req-c43cbdd5-9f03-4553-94e9-466d9f7cc004 req-f6797485-158c-4127-8481-ec03f6bfe4c5 service nova] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Received unexpected event network-vif-plugged-65a91c43-8ff0-4e66-8f18-2de456875968 for instance with vm_state building and task_state spawning. [ 871.256621] env[62965]: DEBUG nova.compute.manager [req-c43cbdd5-9f03-4553-94e9-466d9f7cc004 req-f6797485-158c-4127-8481-ec03f6bfe4c5 service nova] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Received event network-changed-65a91c43-8ff0-4e66-8f18-2de456875968 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 871.256772] env[62965]: DEBUG nova.compute.manager [req-c43cbdd5-9f03-4553-94e9-466d9f7cc004 req-f6797485-158c-4127-8481-ec03f6bfe4c5 service nova] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Refreshing instance network info cache due to event network-changed-65a91c43-8ff0-4e66-8f18-2de456875968. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 871.256932] env[62965]: DEBUG oslo_concurrency.lockutils [req-c43cbdd5-9f03-4553-94e9-466d9f7cc004 req-f6797485-158c-4127-8481-ec03f6bfe4c5 service nova] Acquiring lock "refresh_cache-6def4a38-d3b8-4724-97a1-1859b239d900" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.261322] env[62965]: DEBUG oslo_vmware.api [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1389867, 'name': PowerOnVM_Task, 'duration_secs': 0.681583} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.261560] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 871.261765] env[62965]: INFO nova.compute.manager [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Took 8.18 seconds to spawn the instance on the hypervisor. [ 871.262132] env[62965]: DEBUG nova.compute.manager [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 871.262743] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3939648d-881c-4374-a869-36484f5d0b1f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.415757] env[62965]: DEBUG oslo_concurrency.lockutils [req-28092d41-da55-45d9-9272-2475cc15e9eb req-e24a2c9d-2c47-4c0b-bb3e-67d5933eb1c7 service nova] Releasing lock "refresh_cache-8c478341-9b50-45de-9ba3-321e802ed1f0" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.489092] env[62965]: DEBUG oslo_vmware.api [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389869, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.545688] env[62965]: DEBUG oslo_vmware.api [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Task: {'id': task-1389870, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087413} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.548713] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 871.549478] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9491d762-2455-4f0d-a895-51ab9b064b1b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.556950] env[62965]: DEBUG oslo_vmware.api [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': task-1389868, 'name': ReconfigVM_Task, 'duration_secs': 0.788642} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.565579] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Reconfigured VM instance instance-00000041 to attach disk [datastore2] d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c/d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 871.574824] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] 924c5cdc-0450-4ced-b920-50e2d5060fd2/924c5cdc-0450-4ced-b920-50e2d5060fd2.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 871.575139] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-01067dd1-9d43-437b-b306-96b63d13e0a2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.576689] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7aa6867f-3a6f-4339-9d7e-29e9e3cdf09c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.597598] env[62965]: DEBUG oslo_vmware.api [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Waiting for the task: (returnval){ [ 871.597598] env[62965]: value = "task-1389871" [ 871.597598] env[62965]: _type = "Task" [ 871.597598] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.599016] env[62965]: DEBUG oslo_vmware.api [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Waiting for the task: (returnval){ [ 871.599016] env[62965]: value = "task-1389872" [ 871.599016] env[62965]: _type = "Task" [ 871.599016] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.610041] env[62965]: DEBUG oslo_vmware.api [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': task-1389871, 'name': Rename_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.612941] env[62965]: DEBUG oslo_vmware.api [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Task: {'id': task-1389872, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.636029] env[62965]: DEBUG nova.network.neutron [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 871.638013] env[62965]: DEBUG nova.network.neutron [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.660935] env[62965]: DEBUG nova.scheduler.client.report [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 871.691766] env[62965]: DEBUG oslo_vmware.api [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]525a7f33-b1ac-30fe-a397-7459aec8e675, 'name': SearchDatastore_Task, 'duration_secs': 0.014384} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.692091] env[62965]: DEBUG oslo_concurrency.lockutils [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.692314] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 871.692542] env[62965]: DEBUG oslo_concurrency.lockutils [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.692685] env[62965]: DEBUG oslo_concurrency.lockutils [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.692859] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 871.693122] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f12942f3-6c96-4e3d-8230-bb730fe5ccd9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.701774] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 871.701998] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 871.702698] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-20ed10dc-2b04-4894-9397-467996d4abb2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.709747] env[62965]: DEBUG oslo_vmware.api [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 871.709747] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]525d953d-cf9b-b12e-4fe7-2ddc72d9b247" [ 871.709747] env[62965]: _type = "Task" [ 871.709747] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.716947] env[62965]: DEBUG oslo_vmware.api [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]525d953d-cf9b-b12e-4fe7-2ddc72d9b247, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.781706] env[62965]: INFO nova.compute.manager [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Took 28.83 seconds to build instance. [ 871.924448] env[62965]: DEBUG nova.network.neutron [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Updating instance_info_cache with network_info: [{"id": "65a91c43-8ff0-4e66-8f18-2de456875968", "address": "fa:16:3e:9c:2c:4d", "network": {"id": "98b821d6-efa4-46c8-81e6-a703bd34cb6a", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1649479803-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a539017d4b84f39a71b6f4098395fbe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c979f78-8597-41f8-b1de-995014032689", "external-id": "nsx-vlan-transportzone-477", "segmentation_id": 477, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65a91c43-8f", "ovs_interfaceid": "65a91c43-8ff0-4e66-8f18-2de456875968", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.989291] env[62965]: DEBUG oslo_vmware.api [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389869, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.112492] env[62965]: DEBUG oslo_vmware.api [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Task: {'id': task-1389872, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.115257] env[62965]: DEBUG oslo_vmware.api [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': task-1389871, 'name': Rename_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.141072] env[62965]: INFO nova.compute.manager [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] [instance: 2b6efa32-fc5b-402c-a551-b6e5c01eca2c] Took 1.04 seconds to deallocate network for instance. [ 872.166031] env[62965]: DEBUG oslo_concurrency.lockutils [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.279s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.166570] env[62965]: DEBUG nova.compute.manager [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 872.169377] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 13.832s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.169613] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.169820] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62965) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 872.170227] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.411s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.171737] env[62965]: INFO nova.compute.claims [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 872.175178] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-795fdc11-b138-4ed1-bb71-fdceb2846e9f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.183573] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-147c2482-30e7-4571-92ed-9c8503291fe8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.198642] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-113dadd4-ed9b-453d-bf90-58b4dda8d386 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.205968] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-092002a2-1714-4a8f-bc79-9cf1fcf76dc5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.241057] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181305MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62965) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 872.241259] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.247171] env[62965]: DEBUG oslo_vmware.api [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]525d953d-cf9b-b12e-4fe7-2ddc72d9b247, 'name': SearchDatastore_Task, 'duration_secs': 0.024248} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.247938] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39bb65c3-d42a-40eb-a018-1fc6abe2d093 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.253922] env[62965]: DEBUG oslo_vmware.api [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 872.253922] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5204afad-0d0f-6152-ddf9-fa48a822b31e" [ 872.253922] env[62965]: _type = "Task" [ 872.253922] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.262491] env[62965]: DEBUG oslo_vmware.api [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5204afad-0d0f-6152-ddf9-fa48a822b31e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.281549] env[62965]: DEBUG oslo_concurrency.lockutils [None req-eff1bc97-f58c-4743-9c67-8301e869b2c3 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Lock "b0d8f38a-86d3-466a-9b83-35145071dc21" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 81.129s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.427063] env[62965]: DEBUG oslo_concurrency.lockutils [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Releasing lock "refresh_cache-6def4a38-d3b8-4724-97a1-1859b239d900" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.427507] env[62965]: DEBUG nova.compute.manager [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Instance network_info: |[{"id": "65a91c43-8ff0-4e66-8f18-2de456875968", "address": "fa:16:3e:9c:2c:4d", "network": {"id": "98b821d6-efa4-46c8-81e6-a703bd34cb6a", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1649479803-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a539017d4b84f39a71b6f4098395fbe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c979f78-8597-41f8-b1de-995014032689", "external-id": "nsx-vlan-transportzone-477", "segmentation_id": 477, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65a91c43-8f", "ovs_interfaceid": "65a91c43-8ff0-4e66-8f18-2de456875968", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 872.427825] env[62965]: DEBUG oslo_concurrency.lockutils [req-c43cbdd5-9f03-4553-94e9-466d9f7cc004 req-f6797485-158c-4127-8481-ec03f6bfe4c5 service nova] Acquired lock "refresh_cache-6def4a38-d3b8-4724-97a1-1859b239d900" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.428010] env[62965]: DEBUG nova.network.neutron [req-c43cbdd5-9f03-4553-94e9-466d9f7cc004 req-f6797485-158c-4127-8481-ec03f6bfe4c5 service nova] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Refreshing network info cache for port 65a91c43-8ff0-4e66-8f18-2de456875968 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 872.429224] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9c:2c:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8c979f78-8597-41f8-b1de-995014032689', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '65a91c43-8ff0-4e66-8f18-2de456875968', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 872.436789] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Creating folder: Project (9a539017d4b84f39a71b6f4098395fbe). Parent ref: group-v294931. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 872.439731] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1cbdd7e7-7749-4e37-b590-9d5d4ca06662 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.450517] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Created folder: Project (9a539017d4b84f39a71b6f4098395fbe) in parent group-v294931. [ 872.450727] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Creating folder: Instances. Parent ref: group-v294976. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 872.450959] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d22354cd-b9ac-473a-9bba-ef9b0b0d3b13 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.459689] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Created folder: Instances in parent group-v294976. [ 872.459875] env[62965]: DEBUG oslo.service.loopingcall [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 872.460121] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 872.460381] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dc8c9e5b-b2b9-413b-8c5b-3800aa35488d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.480690] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 872.480690] env[62965]: value = "task-1389875" [ 872.480690] env[62965]: _type = "Task" [ 872.480690] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.490915] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389875, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.493992] env[62965]: DEBUG oslo_vmware.api [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389869, 'name': CreateSnapshot_Task, 'duration_secs': 1.372448} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.494254] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Created Snapshot of the VM instance {{(pid=62965) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 872.494955] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88191dfc-6410-4d81-8807-a23b9c8bcd3f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.612742] env[62965]: DEBUG oslo_vmware.api [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': task-1389871, 'name': Rename_Task, 'duration_secs': 0.802391} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.615866] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 872.616144] env[62965]: DEBUG oslo_vmware.api [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Task: {'id': task-1389872, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.616364] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1e5bc246-be4a-4f86-8df2-93bd8db4cd64 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.622163] env[62965]: DEBUG oslo_vmware.api [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Waiting for the task: (returnval){ [ 872.622163] env[62965]: value = "task-1389876" [ 872.622163] env[62965]: _type = "Task" [ 872.622163] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.629499] env[62965]: DEBUG oslo_vmware.api [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': task-1389876, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.666099] env[62965]: DEBUG nova.network.neutron [req-c43cbdd5-9f03-4553-94e9-466d9f7cc004 req-f6797485-158c-4127-8481-ec03f6bfe4c5 service nova] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Updated VIF entry in instance network info cache for port 65a91c43-8ff0-4e66-8f18-2de456875968. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 872.666465] env[62965]: DEBUG nova.network.neutron [req-c43cbdd5-9f03-4553-94e9-466d9f7cc004 req-f6797485-158c-4127-8481-ec03f6bfe4c5 service nova] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Updating instance_info_cache with network_info: [{"id": "65a91c43-8ff0-4e66-8f18-2de456875968", "address": "fa:16:3e:9c:2c:4d", "network": {"id": "98b821d6-efa4-46c8-81e6-a703bd34cb6a", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1649479803-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a539017d4b84f39a71b6f4098395fbe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c979f78-8597-41f8-b1de-995014032689", "external-id": "nsx-vlan-transportzone-477", "segmentation_id": 477, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65a91c43-8f", "ovs_interfaceid": "65a91c43-8ff0-4e66-8f18-2de456875968", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.675811] env[62965]: DEBUG nova.compute.utils [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 872.677780] env[62965]: DEBUG nova.compute.manager [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 872.678858] env[62965]: DEBUG nova.network.neutron [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 872.720056] env[62965]: DEBUG nova.policy [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd2db1d35b31d477ca92c4ba7a3c53cd3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd91c99b14018469e8998fd1ccc0fab4f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 872.764873] env[62965]: DEBUG oslo_vmware.api [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5204afad-0d0f-6152-ddf9-fa48a822b31e, 'name': SearchDatastore_Task, 'duration_secs': 0.010697} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.765265] env[62965]: DEBUG oslo_concurrency.lockutils [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.765554] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 8c478341-9b50-45de-9ba3-321e802ed1f0/8c478341-9b50-45de-9ba3-321e802ed1f0.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 872.765826] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-53fe7f55-bbb9-4215-82f8-83a25305eb0d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.771759] env[62965]: DEBUG oslo_vmware.api [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 872.771759] env[62965]: value = "task-1389877" [ 872.771759] env[62965]: _type = "Task" [ 872.771759] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.782214] env[62965]: DEBUG oslo_vmware.api [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1389877, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.993467] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389875, 'name': CreateVM_Task, 'duration_secs': 0.354531} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.993868] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 872.994638] env[62965]: DEBUG oslo_concurrency.lockutils [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.994801] env[62965]: DEBUG oslo_concurrency.lockutils [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.995152] env[62965]: DEBUG oslo_concurrency.lockutils [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 872.995413] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89c407be-75a3-44f5-bf44-0ba2fe8238a9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.003547] env[62965]: DEBUG oslo_vmware.api [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Waiting for the task: (returnval){ [ 873.003547] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]528e513a-a05d-f437-90ad-6e09c7c662dd" [ 873.003547] env[62965]: _type = "Task" [ 873.003547] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.016818] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Creating linked-clone VM from snapshot {{(pid=62965) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 873.017231] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-f995f6a4-fd40-4553-9e0a-c2a79100e8d9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.028525] env[62965]: DEBUG oslo_vmware.api [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]528e513a-a05d-f437-90ad-6e09c7c662dd, 'name': SearchDatastore_Task, 'duration_secs': 0.017607} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.030205] env[62965]: DEBUG oslo_concurrency.lockutils [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.030451] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 873.030701] env[62965]: DEBUG oslo_concurrency.lockutils [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.030833] env[62965]: DEBUG oslo_concurrency.lockutils [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.030994] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 873.031352] env[62965]: DEBUG oslo_vmware.api [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 873.031352] env[62965]: value = "task-1389878" [ 873.031352] env[62965]: _type = "Task" [ 873.031352] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.031673] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dd24d49a-4ee3-4da5-bbe0-fbf108554b16 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.044432] env[62965]: DEBUG oslo_vmware.api [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389878, 'name': CloneVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.047119] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 873.047385] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 873.048245] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-acac7d32-7228-4cd5-90b1-b630a7efdc80 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.054523] env[62965]: DEBUG oslo_vmware.api [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Waiting for the task: (returnval){ [ 873.054523] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52d1f8db-4cfd-d8d5-c32d-1a3924a9a630" [ 873.054523] env[62965]: _type = "Task" [ 873.054523] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.064264] env[62965]: DEBUG oslo_vmware.api [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52d1f8db-4cfd-d8d5-c32d-1a3924a9a630, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.115096] env[62965]: DEBUG oslo_vmware.api [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Task: {'id': task-1389872, 'name': ReconfigVM_Task, 'duration_secs': 1.08802} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.116237] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Reconfigured VM instance instance-00000045 to attach disk [datastore2] 924c5cdc-0450-4ced-b920-50e2d5060fd2/924c5cdc-0450-4ced-b920-50e2d5060fd2.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 873.116237] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-338ad989-401d-488c-b8e7-af6d18115005 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.124492] env[62965]: DEBUG oslo_vmware.api [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Waiting for the task: (returnval){ [ 873.124492] env[62965]: value = "task-1389879" [ 873.124492] env[62965]: _type = "Task" [ 873.124492] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.141469] env[62965]: DEBUG oslo_vmware.api [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Task: {'id': task-1389879, 'name': Rename_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.144934] env[62965]: DEBUG oslo_vmware.api [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': task-1389876, 'name': PowerOnVM_Task, 'duration_secs': 0.480163} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.145306] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 873.145522] env[62965]: DEBUG nova.compute.manager [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 873.146374] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf3b8dbd-21af-4318-90b9-76ac4117d499 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.168986] env[62965]: DEBUG oslo_concurrency.lockutils [req-c43cbdd5-9f03-4553-94e9-466d9f7cc004 req-f6797485-158c-4127-8481-ec03f6bfe4c5 service nova] Releasing lock "refresh_cache-6def4a38-d3b8-4724-97a1-1859b239d900" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.182304] env[62965]: DEBUG nova.compute.manager [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 873.192052] env[62965]: DEBUG nova.network.neutron [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Successfully created port: d2d44211-6856-42ed-8f9f-3b1a58b31349 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 873.194851] env[62965]: INFO nova.scheduler.client.report [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] Deleted allocations for instance 2b6efa32-fc5b-402c-a551-b6e5c01eca2c [ 873.281309] env[62965]: DEBUG oslo_vmware.api [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1389877, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.421253] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-460c6911-75dc-4b98-b061-b7d9ab6d7cc9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.427769] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-803ae4b5-8a1d-4e94-9662-230dab14ca53 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.459807] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2a7b112-a22e-4d80-bd17-a8ad5405090c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.467415] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc555f83-4687-4616-b00b-22035e7fd444 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.480722] env[62965]: DEBUG nova.compute.provider_tree [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 873.546683] env[62965]: DEBUG oslo_vmware.api [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389878, 'name': CloneVM_Task} progress is 94%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.564657] env[62965]: DEBUG oslo_vmware.api [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52d1f8db-4cfd-d8d5-c32d-1a3924a9a630, 'name': SearchDatastore_Task, 'duration_secs': 0.023866} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.565132] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0172133-0bf9-4314-866f-f222b39f0ebd {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.569911] env[62965]: DEBUG oslo_vmware.api [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Waiting for the task: (returnval){ [ 873.569911] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52939cc2-db6a-6269-1730-3f0c027f6abe" [ 873.569911] env[62965]: _type = "Task" [ 873.569911] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.577568] env[62965]: DEBUG oslo_vmware.api [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52939cc2-db6a-6269-1730-3f0c027f6abe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.639862] env[62965]: DEBUG oslo_vmware.api [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Task: {'id': task-1389879, 'name': Rename_Task, 'duration_secs': 0.240017} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.640199] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 873.640517] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9ec6fc24-d6c2-4095-a96b-fb3cb5aae885 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.647524] env[62965]: DEBUG oslo_vmware.api [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Waiting for the task: (returnval){ [ 873.647524] env[62965]: value = "task-1389880" [ 873.647524] env[62965]: _type = "Task" [ 873.647524] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.655810] env[62965]: DEBUG oslo_vmware.api [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Task: {'id': task-1389880, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.673491] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.704638] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5a972f86-4f9a-47df-a9e7-681fde64a02e tempest-InstanceActionsTestJSON-1047241527 tempest-InstanceActionsTestJSON-1047241527-project-member] Lock "2b6efa32-fc5b-402c-a551-b6e5c01eca2c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 124.934s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.782584] env[62965]: DEBUG oslo_vmware.api [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1389877, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.534026} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.782855] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 8c478341-9b50-45de-9ba3-321e802ed1f0/8c478341-9b50-45de-9ba3-321e802ed1f0.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 873.783081] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 873.783330] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6f264f44-94b6-4252-852d-0ae7109061a3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.789937] env[62965]: DEBUG oslo_vmware.api [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 873.789937] env[62965]: value = "task-1389881" [ 873.789937] env[62965]: _type = "Task" [ 873.789937] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.798934] env[62965]: DEBUG oslo_vmware.api [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1389881, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.805390] env[62965]: DEBUG nova.compute.manager [None req-320eb21f-761c-4700-80cc-9f87a0353092 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 873.806222] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be2908f8-dbc5-4be6-9376-284401f7252e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.985974] env[62965]: DEBUG nova.scheduler.client.report [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 874.047769] env[62965]: DEBUG oslo_vmware.api [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389878, 'name': CloneVM_Task} progress is 94%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.080597] env[62965]: DEBUG oslo_vmware.api [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52939cc2-db6a-6269-1730-3f0c027f6abe, 'name': SearchDatastore_Task, 'duration_secs': 0.011767} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.080923] env[62965]: DEBUG oslo_concurrency.lockutils [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.081198] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] 6def4a38-d3b8-4724-97a1-1859b239d900/6def4a38-d3b8-4724-97a1-1859b239d900.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 874.081454] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2c19a854-9338-40d7-a682-6ba66724855b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.088151] env[62965]: DEBUG oslo_vmware.api [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Waiting for the task: (returnval){ [ 874.088151] env[62965]: value = "task-1389882" [ 874.088151] env[62965]: _type = "Task" [ 874.088151] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.095850] env[62965]: DEBUG oslo_vmware.api [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1389882, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.158136] env[62965]: DEBUG oslo_vmware.api [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Task: {'id': task-1389880, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.195516] env[62965]: DEBUG nova.compute.manager [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 874.217523] env[62965]: DEBUG nova.virt.hardware [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 874.217776] env[62965]: DEBUG nova.virt.hardware [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 874.217930] env[62965]: DEBUG nova.virt.hardware [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 874.218131] env[62965]: DEBUG nova.virt.hardware [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 874.218277] env[62965]: DEBUG nova.virt.hardware [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 874.218436] env[62965]: DEBUG nova.virt.hardware [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 874.218647] env[62965]: DEBUG nova.virt.hardware [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 874.218808] env[62965]: DEBUG nova.virt.hardware [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 874.218968] env[62965]: DEBUG nova.virt.hardware [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 874.219147] env[62965]: DEBUG nova.virt.hardware [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 874.219321] env[62965]: DEBUG nova.virt.hardware [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 874.220271] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dfcf78d-25fb-4312-94d9-5d181590eb0e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.228277] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ffea084-d70f-47a4-9ce9-117f5d79f4cb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.300716] env[62965]: DEBUG oslo_vmware.api [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1389881, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093803} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.301079] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 874.302194] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f00e4e08-82bf-4a09-8286-b6767e7f18ad {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.321926] env[62965]: INFO nova.compute.manager [None req-320eb21f-761c-4700-80cc-9f87a0353092 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] instance snapshotting [ 874.334116] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] 8c478341-9b50-45de-9ba3-321e802ed1f0/8c478341-9b50-45de-9ba3-321e802ed1f0.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 874.334921] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f9561afa-1f32-4383-9dbd-4359b014b85a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.352463] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90af1dc8-f5f5-4ece-9488-7be530b277e1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.373564] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-432900f0-7983-41b4-9929-ad14f8dac22b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.376432] env[62965]: DEBUG oslo_vmware.api [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 874.376432] env[62965]: value = "task-1389883" [ 874.376432] env[62965]: _type = "Task" [ 874.376432] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.389907] env[62965]: DEBUG oslo_vmware.api [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1389883, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.492351] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.321s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.492351] env[62965]: DEBUG nova.compute.manager [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 874.495115] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.521s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.496489] env[62965]: INFO nova.compute.claims [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 874.545673] env[62965]: DEBUG oslo_concurrency.lockutils [None req-95029734-558f-483b-a8bc-93b5504885ef tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Acquiring lock "d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.545946] env[62965]: DEBUG oslo_concurrency.lockutils [None req-95029734-558f-483b-a8bc-93b5504885ef tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Lock "d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.546170] env[62965]: DEBUG oslo_concurrency.lockutils [None req-95029734-558f-483b-a8bc-93b5504885ef tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Acquiring lock "d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.546353] env[62965]: DEBUG oslo_concurrency.lockutils [None req-95029734-558f-483b-a8bc-93b5504885ef tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Lock "d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.546516] env[62965]: DEBUG oslo_concurrency.lockutils [None req-95029734-558f-483b-a8bc-93b5504885ef tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Lock "d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.548545] env[62965]: INFO nova.compute.manager [None req-95029734-558f-483b-a8bc-93b5504885ef tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Terminating instance [ 874.551904] env[62965]: DEBUG oslo_vmware.api [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389878, 'name': CloneVM_Task} progress is 95%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.598922] env[62965]: DEBUG oslo_vmware.api [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1389882, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.657778] env[62965]: DEBUG oslo_vmware.api [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Task: {'id': task-1389880, 'name': PowerOnVM_Task, 'duration_secs': 0.765515} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.658090] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 874.658304] env[62965]: INFO nova.compute.manager [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Took 9.22 seconds to spawn the instance on the hypervisor. [ 874.658486] env[62965]: DEBUG nova.compute.manager [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 874.659354] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ff2f511-730b-4c91-895f-66ab3602423e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.850326] env[62965]: DEBUG nova.compute.manager [req-b33c91a9-d67b-4db8-96a1-0d01820ab9cf req-ee754b10-fe0f-490d-b6c1-9cdd07211b59 service nova] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Received event network-vif-plugged-d2d44211-6856-42ed-8f9f-3b1a58b31349 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 874.850618] env[62965]: DEBUG oslo_concurrency.lockutils [req-b33c91a9-d67b-4db8-96a1-0d01820ab9cf req-ee754b10-fe0f-490d-b6c1-9cdd07211b59 service nova] Acquiring lock "b59d6023-a047-4b32-bd18-1c7cc05f67d7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.850905] env[62965]: DEBUG oslo_concurrency.lockutils [req-b33c91a9-d67b-4db8-96a1-0d01820ab9cf req-ee754b10-fe0f-490d-b6c1-9cdd07211b59 service nova] Lock "b59d6023-a047-4b32-bd18-1c7cc05f67d7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.851455] env[62965]: DEBUG oslo_concurrency.lockutils [req-b33c91a9-d67b-4db8-96a1-0d01820ab9cf req-ee754b10-fe0f-490d-b6c1-9cdd07211b59 service nova] Lock "b59d6023-a047-4b32-bd18-1c7cc05f67d7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.851723] env[62965]: DEBUG nova.compute.manager [req-b33c91a9-d67b-4db8-96a1-0d01820ab9cf req-ee754b10-fe0f-490d-b6c1-9cdd07211b59 service nova] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] No waiting events found dispatching network-vif-plugged-d2d44211-6856-42ed-8f9f-3b1a58b31349 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 874.852016] env[62965]: WARNING nova.compute.manager [req-b33c91a9-d67b-4db8-96a1-0d01820ab9cf req-ee754b10-fe0f-490d-b6c1-9cdd07211b59 service nova] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Received unexpected event network-vif-plugged-d2d44211-6856-42ed-8f9f-3b1a58b31349 for instance with vm_state building and task_state spawning. [ 874.889092] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-320eb21f-761c-4700-80cc-9f87a0353092 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Creating Snapshot of the VM instance {{(pid=62965) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 874.889420] env[62965]: DEBUG oslo_vmware.api [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1389883, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.889820] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-abb0511b-9ba4-4699-a859-6b508a915611 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.898718] env[62965]: DEBUG oslo_vmware.api [None req-320eb21f-761c-4700-80cc-9f87a0353092 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 874.898718] env[62965]: value = "task-1389884" [ 874.898718] env[62965]: _type = "Task" [ 874.898718] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.909016] env[62965]: DEBUG oslo_vmware.api [None req-320eb21f-761c-4700-80cc-9f87a0353092 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1389884, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.954170] env[62965]: DEBUG nova.network.neutron [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Successfully updated port: d2d44211-6856-42ed-8f9f-3b1a58b31349 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 875.001205] env[62965]: DEBUG nova.compute.utils [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 875.005141] env[62965]: DEBUG nova.compute.manager [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 875.005141] env[62965]: DEBUG nova.network.neutron [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 875.046712] env[62965]: DEBUG oslo_vmware.api [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389878, 'name': CloneVM_Task, 'duration_secs': 1.798159} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.048306] env[62965]: DEBUG nova.policy [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '97ee4fdb7cdb45829c7fe694e883da8c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1ed71548c085499981fa1b102d660368', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 875.049813] env[62965]: INFO nova.virt.vmwareapi.vmops [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Created linked-clone VM from snapshot [ 875.050917] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6419582-e89d-4f8c-ac7e-fb286d91899a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.054228] env[62965]: DEBUG oslo_concurrency.lockutils [None req-95029734-558f-483b-a8bc-93b5504885ef tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Acquiring lock "refresh_cache-d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.054228] env[62965]: DEBUG oslo_concurrency.lockutils [None req-95029734-558f-483b-a8bc-93b5504885ef tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Acquired lock "refresh_cache-d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.054349] env[62965]: DEBUG nova.network.neutron [None req-95029734-558f-483b-a8bc-93b5504885ef tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 875.062379] env[62965]: DEBUG nova.virt.vmwareapi.images [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Uploading image 7e436a80-d1ff-4c42-847a-9538961b1a6a {{(pid=62965) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 875.090244] env[62965]: DEBUG oslo_vmware.rw_handles [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 875.090244] env[62965]: value = "vm-294979" [ 875.090244] env[62965]: _type = "VirtualMachine" [ 875.090244] env[62965]: }. {{(pid=62965) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 875.090587] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-74a5c2e3-6cf3-4b65-bbe8-5241a66bd683 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.105019] env[62965]: DEBUG oslo_vmware.api [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1389882, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.924444} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.105019] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] 6def4a38-d3b8-4724-97a1-1859b239d900/6def4a38-d3b8-4724-97a1-1859b239d900.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 875.105019] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 875.105019] env[62965]: DEBUG oslo_vmware.rw_handles [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lease: (returnval){ [ 875.105019] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52b3f64c-9a5c-5a59-c7c7-61ad97c2782b" [ 875.105019] env[62965]: _type = "HttpNfcLease" [ 875.105019] env[62965]: } obtained for exporting VM: (result){ [ 875.105019] env[62965]: value = "vm-294979" [ 875.105019] env[62965]: _type = "VirtualMachine" [ 875.105019] env[62965]: }. {{(pid=62965) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 875.105509] env[62965]: DEBUG oslo_vmware.api [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the lease: (returnval){ [ 875.105509] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52b3f64c-9a5c-5a59-c7c7-61ad97c2782b" [ 875.105509] env[62965]: _type = "HttpNfcLease" [ 875.105509] env[62965]: } to be ready. {{(pid=62965) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 875.105509] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6a3b758c-65be-4232-ac81-c5b95cf01cd1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.112059] env[62965]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 875.112059] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52b3f64c-9a5c-5a59-c7c7-61ad97c2782b" [ 875.112059] env[62965]: _type = "HttpNfcLease" [ 875.112059] env[62965]: } is initializing. {{(pid=62965) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 875.114206] env[62965]: DEBUG oslo_vmware.api [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Waiting for the task: (returnval){ [ 875.114206] env[62965]: value = "task-1389886" [ 875.114206] env[62965]: _type = "Task" [ 875.114206] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.122087] env[62965]: DEBUG oslo_vmware.api [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1389886, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.179015] env[62965]: INFO nova.compute.manager [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Took 30.49 seconds to build instance. [ 875.391593] env[62965]: DEBUG oslo_vmware.api [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1389883, 'name': ReconfigVM_Task, 'duration_secs': 0.570526} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.391593] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Reconfigured VM instance instance-00000046 to attach disk [datastore2] 8c478341-9b50-45de-9ba3-321e802ed1f0/8c478341-9b50-45de-9ba3-321e802ed1f0.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 875.391593] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1d5533bc-899c-42e1-a505-d1118785c089 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.397163] env[62965]: DEBUG oslo_vmware.api [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 875.397163] env[62965]: value = "task-1389887" [ 875.397163] env[62965]: _type = "Task" [ 875.397163] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.409129] env[62965]: DEBUG oslo_vmware.api [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1389887, 'name': Rename_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.411292] env[62965]: DEBUG oslo_vmware.api [None req-320eb21f-761c-4700-80cc-9f87a0353092 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1389884, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.437646] env[62965]: DEBUG nova.network.neutron [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Successfully created port: 7faa0a61-99a2-4ae4-b60a-ef2eba7aef31 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 875.457968] env[62965]: DEBUG oslo_concurrency.lockutils [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquiring lock "refresh_cache-b59d6023-a047-4b32-bd18-1c7cc05f67d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.457968] env[62965]: DEBUG oslo_concurrency.lockutils [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquired lock "refresh_cache-b59d6023-a047-4b32-bd18-1c7cc05f67d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.457968] env[62965]: DEBUG nova.network.neutron [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 875.511755] env[62965]: DEBUG nova.compute.manager [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 875.594840] env[62965]: DEBUG nova.network.neutron [None req-95029734-558f-483b-a8bc-93b5504885ef tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 875.618131] env[62965]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 875.618131] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52b3f64c-9a5c-5a59-c7c7-61ad97c2782b" [ 875.618131] env[62965]: _type = "HttpNfcLease" [ 875.618131] env[62965]: } is ready. {{(pid=62965) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 875.621948] env[62965]: DEBUG oslo_vmware.rw_handles [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 875.621948] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52b3f64c-9a5c-5a59-c7c7-61ad97c2782b" [ 875.621948] env[62965]: _type = "HttpNfcLease" [ 875.621948] env[62965]: }. {{(pid=62965) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 875.622829] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91a4985f-69ce-44d4-b0fa-83e906f8bb67 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.632717] env[62965]: DEBUG oslo_vmware.api [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1389886, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.206216} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.639167] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 875.640260] env[62965]: DEBUG oslo_vmware.rw_handles [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52be7c0e-3874-7f2c-a856-0fff94f43cde/disk-0.vmdk from lease info. {{(pid=62965) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 875.640455] env[62965]: DEBUG oslo_vmware.rw_handles [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52be7c0e-3874-7f2c-a856-0fff94f43cde/disk-0.vmdk for reading. {{(pid=62965) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 875.643625] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcb9821d-8ea8-431f-a59b-6a52d484cd71 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.718267] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1b2724c8-487b-4ae0-840e-d02ef2c5713e tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Lock "924c5cdc-0450-4ced-b920-50e2d5060fd2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 67.016s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.732769] env[62965]: DEBUG nova.network.neutron [None req-95029734-558f-483b-a8bc-93b5504885ef tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.743642] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] 6def4a38-d3b8-4724-97a1-1859b239d900/6def4a38-d3b8-4724-97a1-1859b239d900.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 875.746136] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec1c6570-2ab2-4fba-98e1-3ad03510a123 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.771240] env[62965]: DEBUG oslo_vmware.api [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Waiting for the task: (returnval){ [ 875.771240] env[62965]: value = "task-1389888" [ 875.771240] env[62965]: _type = "Task" [ 875.771240] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.786281] env[62965]: DEBUG oslo_vmware.api [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1389888, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.792921] env[62965]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-7f956892-f76a-4067-8b1e-bb913971518a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.876167] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00b23fe4-ac02-48ce-bfc7-13e338198e7d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.884322] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c328e32-0677-4fa6-91c0-1e734935c580 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.923176] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c683dfd-3d1c-46e9-8210-5067cebe66b3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.931583] env[62965]: DEBUG oslo_vmware.api [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1389887, 'name': Rename_Task, 'duration_secs': 0.265833} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.937146] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 875.937146] env[62965]: DEBUG oslo_vmware.api [None req-320eb21f-761c-4700-80cc-9f87a0353092 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1389884, 'name': CreateSnapshot_Task, 'duration_secs': 0.953529} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.937260] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fcd79bbe-ad00-4e3b-9af3-0610c94bb7c6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.939985] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c96db4c7-56f7-4ed4-9c96-f56816cab1fb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.944378] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-320eb21f-761c-4700-80cc-9f87a0353092 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Created Snapshot of the VM instance {{(pid=62965) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 875.945131] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1474612f-63a3-4002-a487-cbfd64604e47 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.958635] env[62965]: DEBUG nova.compute.provider_tree [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 875.970018] env[62965]: DEBUG oslo_vmware.api [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 875.970018] env[62965]: value = "task-1389889" [ 875.970018] env[62965]: _type = "Task" [ 875.970018] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.984355] env[62965]: DEBUG oslo_vmware.api [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1389889, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.012554] env[62965]: DEBUG nova.network.neutron [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 876.192174] env[62965]: DEBUG nova.network.neutron [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Updating instance_info_cache with network_info: [{"id": "d2d44211-6856-42ed-8f9f-3b1a58b31349", "address": "fa:16:3e:3b:f0:e9", "network": {"id": "49dd42b0-6397-4fdd-a319-ed920c72a5af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-373172566-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d91c99b14018469e8998fd1ccc0fab4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2d44211-68", "ovs_interfaceid": "d2d44211-6856-42ed-8f9f-3b1a58b31349", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.247115] env[62965]: DEBUG oslo_concurrency.lockutils [None req-95029734-558f-483b-a8bc-93b5504885ef tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Releasing lock "refresh_cache-d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.249682] env[62965]: DEBUG nova.compute.manager [None req-95029734-558f-483b-a8bc-93b5504885ef tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 876.249808] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-95029734-558f-483b-a8bc-93b5504885ef tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 876.250713] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-493dd2ed-6cf5-4fbe-92e3-f7b709e9444f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.270471] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-95029734-558f-483b-a8bc-93b5504885ef tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 876.270853] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-00cb68c9-f6a0-4330-a365-e9c523a4b694 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.285276] env[62965]: DEBUG oslo_vmware.api [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1389888, 'name': ReconfigVM_Task, 'duration_secs': 0.5083} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.288592] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Reconfigured VM instance instance-00000047 to attach disk [datastore1] 6def4a38-d3b8-4724-97a1-1859b239d900/6def4a38-d3b8-4724-97a1-1859b239d900.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 876.290317] env[62965]: DEBUG oslo_vmware.api [None req-95029734-558f-483b-a8bc-93b5504885ef tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Waiting for the task: (returnval){ [ 876.290317] env[62965]: value = "task-1389890" [ 876.290317] env[62965]: _type = "Task" [ 876.290317] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.290606] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-88c9727b-e3bf-41ff-b880-6c14fd04a265 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.301068] env[62965]: DEBUG oslo_vmware.api [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Waiting for the task: (returnval){ [ 876.301068] env[62965]: value = "task-1389891" [ 876.301068] env[62965]: _type = "Task" [ 876.301068] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.307015] env[62965]: DEBUG oslo_vmware.api [None req-95029734-558f-483b-a8bc-93b5504885ef tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': task-1389890, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.319885] env[62965]: DEBUG oslo_vmware.api [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1389891, 'name': Rename_Task} progress is 10%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.462675] env[62965]: DEBUG nova.scheduler.client.report [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 876.483323] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-320eb21f-761c-4700-80cc-9f87a0353092 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Creating linked-clone VM from snapshot {{(pid=62965) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 876.484302] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-99eeb2d4-426d-4121-8b63-b9fb0aa30606 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.494052] env[62965]: DEBUG oslo_vmware.api [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1389889, 'name': PowerOnVM_Task} progress is 71%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.495794] env[62965]: DEBUG oslo_vmware.api [None req-320eb21f-761c-4700-80cc-9f87a0353092 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 876.495794] env[62965]: value = "task-1389892" [ 876.495794] env[62965]: _type = "Task" [ 876.495794] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.506105] env[62965]: DEBUG oslo_vmware.api [None req-320eb21f-761c-4700-80cc-9f87a0353092 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1389892, 'name': CloneVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.522228] env[62965]: DEBUG nova.compute.manager [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 876.552972] env[62965]: DEBUG nova.virt.hardware [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 876.553380] env[62965]: DEBUG nova.virt.hardware [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 876.553561] env[62965]: DEBUG nova.virt.hardware [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 876.553882] env[62965]: DEBUG nova.virt.hardware [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 876.554062] env[62965]: DEBUG nova.virt.hardware [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 876.554226] env[62965]: DEBUG nova.virt.hardware [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 876.554712] env[62965]: DEBUG nova.virt.hardware [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 876.554712] env[62965]: DEBUG nova.virt.hardware [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 876.554933] env[62965]: DEBUG nova.virt.hardware [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 876.555391] env[62965]: DEBUG nova.virt.hardware [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 876.555626] env[62965]: DEBUG nova.virt.hardware [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 876.557380] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67cf2bf8-795a-4bb4-b7dc-46625043d4eb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.567257] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d819d572-dce0-4c20-8bff-d9772399b57a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.695694] env[62965]: DEBUG oslo_concurrency.lockutils [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Releasing lock "refresh_cache-b59d6023-a047-4b32-bd18-1c7cc05f67d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.696037] env[62965]: DEBUG nova.compute.manager [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Instance network_info: |[{"id": "d2d44211-6856-42ed-8f9f-3b1a58b31349", "address": "fa:16:3e:3b:f0:e9", "network": {"id": "49dd42b0-6397-4fdd-a319-ed920c72a5af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-373172566-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d91c99b14018469e8998fd1ccc0fab4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2d44211-68", "ovs_interfaceid": "d2d44211-6856-42ed-8f9f-3b1a58b31349", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 876.697572] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3b:f0:e9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f77ff7a1-209c-4f3f-b2a0-fd817741e739', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd2d44211-6856-42ed-8f9f-3b1a58b31349', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 876.712577] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Creating folder: Project (d91c99b14018469e8998fd1ccc0fab4f). Parent ref: group-v294931. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 876.712823] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-668190e9-d666-4393-8773-de7383d6a559 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.724871] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Created folder: Project (d91c99b14018469e8998fd1ccc0fab4f) in parent group-v294931. [ 876.725113] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Creating folder: Instances. Parent ref: group-v294982. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 876.725378] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d0c8326e-1a7d-4cf5-9aa0-ab2d71dcb270 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.734806] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Created folder: Instances in parent group-v294982. [ 876.735071] env[62965]: DEBUG oslo.service.loopingcall [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 876.735279] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 876.735488] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3444f901-c0ca-4a82-b083-a143069fd630 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.757457] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 876.757457] env[62965]: value = "task-1389895" [ 876.757457] env[62965]: _type = "Task" [ 876.757457] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.767028] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389895, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.802783] env[62965]: DEBUG oslo_vmware.api [None req-95029734-558f-483b-a8bc-93b5504885ef tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': task-1389890, 'name': PowerOffVM_Task, 'duration_secs': 0.330306} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.803197] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-95029734-558f-483b-a8bc-93b5504885ef tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 876.803397] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-95029734-558f-483b-a8bc-93b5504885ef tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 876.804089] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fb226c67-325f-4262-a9d1-c00c90ae9660 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.817137] env[62965]: DEBUG oslo_vmware.api [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1389891, 'name': Rename_Task, 'duration_secs': 0.23105} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.820935] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 876.820935] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-886d737b-6422-4770-bfec-db376e606aab {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.824538] env[62965]: DEBUG oslo_vmware.api [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Waiting for the task: (returnval){ [ 876.824538] env[62965]: value = "task-1389897" [ 876.824538] env[62965]: _type = "Task" [ 876.824538] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.834264] env[62965]: DEBUG oslo_vmware.api [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1389897, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.835436] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-95029734-558f-483b-a8bc-93b5504885ef tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 876.837781] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-95029734-558f-483b-a8bc-93b5504885ef tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Deleting contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 876.837781] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-95029734-558f-483b-a8bc-93b5504885ef tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Deleting the datastore file [datastore2] d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 876.837781] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9d387677-8f82-4f64-864d-13f77c8445da {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.843999] env[62965]: DEBUG oslo_vmware.api [None req-95029734-558f-483b-a8bc-93b5504885ef tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Waiting for the task: (returnval){ [ 876.843999] env[62965]: value = "task-1389898" [ 876.843999] env[62965]: _type = "Task" [ 876.843999] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.856628] env[62965]: DEBUG oslo_vmware.api [None req-95029734-558f-483b-a8bc-93b5504885ef tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': task-1389898, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.964929] env[62965]: DEBUG nova.compute.manager [req-85d9ce19-9d69-42d9-9c29-d986af3831c6 req-83a723d2-1243-4f69-9670-f24debaccf79 service nova] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Received event network-changed-d2d44211-6856-42ed-8f9f-3b1a58b31349 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 876.965294] env[62965]: DEBUG nova.compute.manager [req-85d9ce19-9d69-42d9-9c29-d986af3831c6 req-83a723d2-1243-4f69-9670-f24debaccf79 service nova] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Refreshing instance network info cache due to event network-changed-d2d44211-6856-42ed-8f9f-3b1a58b31349. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 876.965589] env[62965]: DEBUG oslo_concurrency.lockutils [req-85d9ce19-9d69-42d9-9c29-d986af3831c6 req-83a723d2-1243-4f69-9670-f24debaccf79 service nova] Acquiring lock "refresh_cache-b59d6023-a047-4b32-bd18-1c7cc05f67d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.965793] env[62965]: DEBUG oslo_concurrency.lockutils [req-85d9ce19-9d69-42d9-9c29-d986af3831c6 req-83a723d2-1243-4f69-9670-f24debaccf79 service nova] Acquired lock "refresh_cache-b59d6023-a047-4b32-bd18-1c7cc05f67d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.966020] env[62965]: DEBUG nova.network.neutron [req-85d9ce19-9d69-42d9-9c29-d986af3831c6 req-83a723d2-1243-4f69-9670-f24debaccf79 service nova] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Refreshing network info cache for port d2d44211-6856-42ed-8f9f-3b1a58b31349 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 876.969451] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.474s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.970166] env[62965]: DEBUG nova.compute.manager [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 876.979994] env[62965]: DEBUG oslo_concurrency.lockutils [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.619s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.981573] env[62965]: INFO nova.compute.claims [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 876.999048] env[62965]: DEBUG oslo_vmware.api [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1389889, 'name': PowerOnVM_Task, 'duration_secs': 0.892512} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.003213] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 877.003522] env[62965]: INFO nova.compute.manager [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Took 9.23 seconds to spawn the instance on the hypervisor. [ 877.003784] env[62965]: DEBUG nova.compute.manager [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 877.004988] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc8278b2-9b0e-45d5-85cb-a9926cedf512 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.016515] env[62965]: DEBUG oslo_vmware.api [None req-320eb21f-761c-4700-80cc-9f87a0353092 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1389892, 'name': CloneVM_Task} progress is 94%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.270951] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389895, 'name': CreateVM_Task} progress is 99%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.336717] env[62965]: DEBUG oslo_vmware.api [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1389897, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.357888] env[62965]: DEBUG oslo_vmware.api [None req-95029734-558f-483b-a8bc-93b5504885ef tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Task: {'id': task-1389898, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.189949} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.358564] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-95029734-558f-483b-a8bc-93b5504885ef tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 877.358909] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-95029734-558f-483b-a8bc-93b5504885ef tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Deleted contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 877.359249] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-95029734-558f-483b-a8bc-93b5504885ef tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 877.360093] env[62965]: INFO nova.compute.manager [None req-95029734-558f-483b-a8bc-93b5504885ef tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Took 1.11 seconds to destroy the instance on the hypervisor. [ 877.360093] env[62965]: DEBUG oslo.service.loopingcall [None req-95029734-558f-483b-a8bc-93b5504885ef tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 877.360093] env[62965]: DEBUG nova.compute.manager [-] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 877.360248] env[62965]: DEBUG nova.network.neutron [-] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 877.386467] env[62965]: DEBUG nova.network.neutron [-] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 877.481801] env[62965]: DEBUG nova.compute.utils [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 877.485808] env[62965]: DEBUG nova.compute.manager [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 877.486091] env[62965]: DEBUG nova.network.neutron [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 877.510335] env[62965]: DEBUG oslo_vmware.api [None req-320eb21f-761c-4700-80cc-9f87a0353092 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1389892, 'name': CloneVM_Task} progress is 94%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.532083] env[62965]: INFO nova.compute.manager [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Took 29.19 seconds to build instance. [ 877.558711] env[62965]: DEBUG nova.network.neutron [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Successfully updated port: 7faa0a61-99a2-4ae4-b60a-ef2eba7aef31 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 877.597097] env[62965]: DEBUG nova.policy [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f3c1bb4b1cd34098acea01a7656c0a57', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '89fd5275b2064288b44e79365f6f0271', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 877.776079] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389895, 'name': CreateVM_Task, 'duration_secs': 0.544362} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.776367] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 877.781110] env[62965]: DEBUG oslo_concurrency.lockutils [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.781319] env[62965]: DEBUG oslo_concurrency.lockutils [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.781665] env[62965]: DEBUG oslo_concurrency.lockutils [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 877.782105] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-afc2ada5-dd80-4408-9eda-74e16e7382c4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.787839] env[62965]: DEBUG oslo_vmware.api [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 877.787839] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]529872d1-2fc5-1ad1-8589-0e27a4324483" [ 877.787839] env[62965]: _type = "Task" [ 877.787839] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.799118] env[62965]: DEBUG oslo_vmware.api [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]529872d1-2fc5-1ad1-8589-0e27a4324483, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.834354] env[62965]: DEBUG oslo_vmware.api [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1389897, 'name': PowerOnVM_Task, 'duration_secs': 0.61327} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.837288] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 877.837288] env[62965]: INFO nova.compute.manager [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Took 7.77 seconds to spawn the instance on the hypervisor. [ 877.837288] env[62965]: DEBUG nova.compute.manager [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 877.837288] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bb9f31e-3ac4-406a-979b-255ecd5f5747 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.889827] env[62965]: DEBUG nova.network.neutron [-] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.941525] env[62965]: DEBUG nova.network.neutron [req-85d9ce19-9d69-42d9-9c29-d986af3831c6 req-83a723d2-1243-4f69-9670-f24debaccf79 service nova] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Updated VIF entry in instance network info cache for port d2d44211-6856-42ed-8f9f-3b1a58b31349. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 877.941975] env[62965]: DEBUG nova.network.neutron [req-85d9ce19-9d69-42d9-9c29-d986af3831c6 req-83a723d2-1243-4f69-9670-f24debaccf79 service nova] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Updating instance_info_cache with network_info: [{"id": "d2d44211-6856-42ed-8f9f-3b1a58b31349", "address": "fa:16:3e:3b:f0:e9", "network": {"id": "49dd42b0-6397-4fdd-a319-ed920c72a5af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-373172566-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d91c99b14018469e8998fd1ccc0fab4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2d44211-68", "ovs_interfaceid": "d2d44211-6856-42ed-8f9f-3b1a58b31349", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.987132] env[62965]: DEBUG nova.compute.manager [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 878.009070] env[62965]: DEBUG oslo_vmware.api [None req-320eb21f-761c-4700-80cc-9f87a0353092 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1389892, 'name': CloneVM_Task} progress is 100%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.035315] env[62965]: DEBUG oslo_concurrency.lockutils [None req-89612c48-b1ad-4546-8e27-77164f416059 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "8c478341-9b50-45de-9ba3-321e802ed1f0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 65.015s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.061648] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "refresh_cache-bd36601b-5a70-4a29-8ef1-d1e925f41de7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.061792] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquired lock "refresh_cache-bd36601b-5a70-4a29-8ef1-d1e925f41de7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.061982] env[62965]: DEBUG nova.network.neutron [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 878.065902] env[62965]: DEBUG nova.network.neutron [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Successfully created port: 6ffa156f-6a68-424e-b324-660892b9ad22 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 878.212021] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d0b29aa-6b28-455d-9ed6-5ac3e398f0a0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.221288] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3456bbcc-6904-498e-8758-9fe73eed910a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.254359] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7bb8c3c-1a13-4ddf-8d16-d7acc0593f89 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.262892] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3589f758-8089-4836-ab42-8e72776d99d6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.277150] env[62965]: DEBUG nova.compute.provider_tree [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 878.299755] env[62965]: DEBUG oslo_vmware.api [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]529872d1-2fc5-1ad1-8589-0e27a4324483, 'name': SearchDatastore_Task, 'duration_secs': 0.019742} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.300684] env[62965]: DEBUG oslo_concurrency.lockutils [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.300684] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 878.300684] env[62965]: DEBUG oslo_concurrency.lockutils [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.300928] env[62965]: DEBUG oslo_concurrency.lockutils [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.301058] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 878.301346] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c03ec818-a3f3-4968-b5cd-f870634f275c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.311320] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 878.311493] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 878.312759] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-44ade19c-614c-40b3-b854-bb64d1e946cb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.317995] env[62965]: DEBUG oslo_vmware.api [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 878.317995] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]522fe193-1491-e955-a783-b105c439da55" [ 878.317995] env[62965]: _type = "Task" [ 878.317995] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.327384] env[62965]: DEBUG oslo_vmware.api [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]522fe193-1491-e955-a783-b105c439da55, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.359054] env[62965]: INFO nova.compute.manager [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Took 28.26 seconds to build instance. [ 878.393017] env[62965]: INFO nova.compute.manager [-] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Took 1.03 seconds to deallocate network for instance. [ 878.444757] env[62965]: DEBUG oslo_concurrency.lockutils [req-85d9ce19-9d69-42d9-9c29-d986af3831c6 req-83a723d2-1243-4f69-9670-f24debaccf79 service nova] Releasing lock "refresh_cache-b59d6023-a047-4b32-bd18-1c7cc05f67d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.516396] env[62965]: DEBUG oslo_vmware.api [None req-320eb21f-761c-4700-80cc-9f87a0353092 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1389892, 'name': CloneVM_Task, 'duration_secs': 1.555377} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.516683] env[62965]: INFO nova.virt.vmwareapi.vmops [None req-320eb21f-761c-4700-80cc-9f87a0353092 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Created linked-clone VM from snapshot [ 878.517643] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4f0eb16-07e5-4ccd-a266-6e7562a1d1b3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.529025] env[62965]: DEBUG nova.virt.vmwareapi.images [None req-320eb21f-761c-4700-80cc-9f87a0353092 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Uploading image ed859b32-0dd3-4eca-8639-006d78224e79 {{(pid=62965) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 878.553125] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-320eb21f-761c-4700-80cc-9f87a0353092 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Destroying the VM {{(pid=62965) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 878.553125] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-a2dbde2d-b431-4045-81d4-31df1b49b1e9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.559438] env[62965]: DEBUG oslo_vmware.api [None req-320eb21f-761c-4700-80cc-9f87a0353092 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 878.559438] env[62965]: value = "task-1389899" [ 878.559438] env[62965]: _type = "Task" [ 878.559438] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.570350] env[62965]: DEBUG oslo_vmware.api [None req-320eb21f-761c-4700-80cc-9f87a0353092 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1389899, 'name': Destroy_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.607474] env[62965]: DEBUG nova.network.neutron [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 878.752511] env[62965]: DEBUG nova.network.neutron [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Updating instance_info_cache with network_info: [{"id": "7faa0a61-99a2-4ae4-b60a-ef2eba7aef31", "address": "fa:16:3e:ec:07:ba", "network": {"id": "a0a5e605-d580-4b7b-b3c0-5c7395bd5a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1679618017-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ed71548c085499981fa1b102d660368", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7faa0a61-99", "ovs_interfaceid": "7faa0a61-99a2-4ae4-b60a-ef2eba7aef31", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.783026] env[62965]: DEBUG nova.scheduler.client.report [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 878.833278] env[62965]: DEBUG oslo_vmware.api [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]522fe193-1491-e955-a783-b105c439da55, 'name': SearchDatastore_Task, 'duration_secs': 0.012117} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.833882] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca5907bf-cbb3-4459-b3c2-bf3262b16984 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.841372] env[62965]: DEBUG oslo_vmware.api [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 878.841372] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]527bf649-d03d-cadb-cd52-e3245c112ae5" [ 878.841372] env[62965]: _type = "Task" [ 878.841372] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.858169] env[62965]: DEBUG oslo_vmware.api [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]527bf649-d03d-cadb-cd52-e3245c112ae5, 'name': SearchDatastore_Task, 'duration_secs': 0.01235} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.859618] env[62965]: DEBUG oslo_concurrency.lockutils [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.860079] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] b59d6023-a047-4b32-bd18-1c7cc05f67d7/b59d6023-a047-4b32-bd18-1c7cc05f67d7.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 878.860688] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1bd6ba0a-71c1-4596-8075-70ec8b36e340 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.865268] env[62965]: DEBUG oslo_concurrency.lockutils [None req-adaa1f76-7345-47b5-bed5-c6a5ab27e380 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "6def4a38-d3b8-4724-97a1-1859b239d900" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.501s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.870798] env[62965]: DEBUG oslo_vmware.api [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 878.870798] env[62965]: value = "task-1389900" [ 878.870798] env[62965]: _type = "Task" [ 878.870798] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.880230] env[62965]: DEBUG oslo_vmware.api [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1389900, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.901105] env[62965]: DEBUG oslo_concurrency.lockutils [None req-95029734-558f-483b-a8bc-93b5504885ef tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.996438] env[62965]: DEBUG nova.compute.manager [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 879.005093] env[62965]: DEBUG nova.compute.manager [req-03a27643-c7fd-4a18-ba55-8c1c77a19ff4 req-90985482-7e18-47b1-9e1f-9a61cbd13fac service nova] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Received event network-changed-80afc0d2-0d83-4374-a4c0-792eabf66c40 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 879.005452] env[62965]: DEBUG nova.compute.manager [req-03a27643-c7fd-4a18-ba55-8c1c77a19ff4 req-90985482-7e18-47b1-9e1f-9a61cbd13fac service nova] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Refreshing instance network info cache due to event network-changed-80afc0d2-0d83-4374-a4c0-792eabf66c40. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 879.006410] env[62965]: DEBUG oslo_concurrency.lockutils [req-03a27643-c7fd-4a18-ba55-8c1c77a19ff4 req-90985482-7e18-47b1-9e1f-9a61cbd13fac service nova] Acquiring lock "refresh_cache-924c5cdc-0450-4ced-b920-50e2d5060fd2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.006410] env[62965]: DEBUG oslo_concurrency.lockutils [req-03a27643-c7fd-4a18-ba55-8c1c77a19ff4 req-90985482-7e18-47b1-9e1f-9a61cbd13fac service nova] Acquired lock "refresh_cache-924c5cdc-0450-4ced-b920-50e2d5060fd2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.006410] env[62965]: DEBUG nova.network.neutron [req-03a27643-c7fd-4a18-ba55-8c1c77a19ff4 req-90985482-7e18-47b1-9e1f-9a61cbd13fac service nova] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Refreshing network info cache for port 80afc0d2-0d83-4374-a4c0-792eabf66c40 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 879.023083] env[62965]: DEBUG nova.virt.hardware [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 879.027262] env[62965]: DEBUG nova.virt.hardware [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 879.027441] env[62965]: DEBUG nova.virt.hardware [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 879.027683] env[62965]: DEBUG nova.virt.hardware [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 879.028888] env[62965]: DEBUG nova.virt.hardware [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 879.028888] env[62965]: DEBUG nova.virt.hardware [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 879.028888] env[62965]: DEBUG nova.virt.hardware [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 879.028888] env[62965]: DEBUG nova.virt.hardware [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 879.028888] env[62965]: DEBUG nova.virt.hardware [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 879.029099] env[62965]: DEBUG nova.virt.hardware [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 879.029131] env[62965]: DEBUG nova.virt.hardware [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 879.030851] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec025dee-a530-4a8f-809d-958a5fe483e9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.041236] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c0f049b-7c19-4767-8ec8-4923b3cd6fb9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.070121] env[62965]: DEBUG oslo_vmware.api [None req-320eb21f-761c-4700-80cc-9f87a0353092 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1389899, 'name': Destroy_Task, 'duration_secs': 0.501834} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.070279] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-320eb21f-761c-4700-80cc-9f87a0353092 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Destroyed the VM [ 879.070980] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-320eb21f-761c-4700-80cc-9f87a0353092 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Deleting Snapshot of the VM instance {{(pid=62965) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 879.070980] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-26b5b7cd-1a39-405d-a267-ca8d97a399cb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.079693] env[62965]: DEBUG oslo_vmware.api [None req-320eb21f-761c-4700-80cc-9f87a0353092 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 879.079693] env[62965]: value = "task-1389901" [ 879.079693] env[62965]: _type = "Task" [ 879.079693] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.091705] env[62965]: DEBUG oslo_vmware.api [None req-320eb21f-761c-4700-80cc-9f87a0353092 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1389901, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.257288] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Releasing lock "refresh_cache-bd36601b-5a70-4a29-8ef1-d1e925f41de7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.257725] env[62965]: DEBUG nova.compute.manager [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Instance network_info: |[{"id": "7faa0a61-99a2-4ae4-b60a-ef2eba7aef31", "address": "fa:16:3e:ec:07:ba", "network": {"id": "a0a5e605-d580-4b7b-b3c0-5c7395bd5a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1679618017-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ed71548c085499981fa1b102d660368", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7faa0a61-99", "ovs_interfaceid": "7faa0a61-99a2-4ae4-b60a-ef2eba7aef31", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 879.258225] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ec:07:ba', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0cd5d325-3053-407e-a4ee-f627e82a23f9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7faa0a61-99a2-4ae4-b60a-ef2eba7aef31', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 879.267096] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Creating folder: Project (1ed71548c085499981fa1b102d660368). Parent ref: group-v294931. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 879.267536] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7b5883f7-fe18-4be0-8e81-794f431ec9f9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.278194] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Created folder: Project (1ed71548c085499981fa1b102d660368) in parent group-v294931. [ 879.278595] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Creating folder: Instances. Parent ref: group-v294985. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 879.278782] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9e9b89cc-5ff3-40e5-bcd4-2f9f758f4f24 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.285530] env[62965]: DEBUG oslo_concurrency.lockutils [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.306s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.286176] env[62965]: DEBUG nova.compute.manager [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 879.292035] env[62965]: DEBUG oslo_concurrency.lockutils [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.453s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.295179] env[62965]: INFO nova.compute.claims [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 879.299706] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Created folder: Instances in parent group-v294985. [ 879.300314] env[62965]: DEBUG oslo.service.loopingcall [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 879.301767] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 879.302377] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ffd05c76-01e2-4ef5-89a1-12312ecc46ee {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.332641] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 879.332641] env[62965]: value = "task-1389904" [ 879.332641] env[62965]: _type = "Task" [ 879.332641] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.347443] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389904, 'name': CreateVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.386382] env[62965]: DEBUG oslo_vmware.api [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1389900, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.595292] env[62965]: DEBUG oslo_vmware.api [None req-320eb21f-761c-4700-80cc-9f87a0353092 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1389901, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.632024] env[62965]: INFO nova.compute.manager [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Rebuilding instance [ 879.694045] env[62965]: DEBUG nova.compute.manager [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 879.694992] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b66a8d5-4573-44c0-a447-26fd98ccc86c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.796580] env[62965]: DEBUG nova.network.neutron [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Successfully updated port: 6ffa156f-6a68-424e-b324-660892b9ad22 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 879.809725] env[62965]: DEBUG nova.compute.utils [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 879.812852] env[62965]: DEBUG nova.compute.manager [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 879.813264] env[62965]: DEBUG nova.network.neutron [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 879.843951] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389904, 'name': CreateVM_Task, 'duration_secs': 0.379982} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.844298] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 879.845106] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.845519] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.846327] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 879.846983] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d71607c1-b158-4d2d-b7d9-576c51e594da {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.852869] env[62965]: DEBUG oslo_vmware.api [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 879.852869] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]520ed17e-91b9-afbb-3cbf-dae27a1b10b0" [ 879.852869] env[62965]: _type = "Task" [ 879.852869] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.862978] env[62965]: DEBUG oslo_vmware.api [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]520ed17e-91b9-afbb-3cbf-dae27a1b10b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.878187] env[62965]: DEBUG nova.policy [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03c0a9a0ec3c46bf883c68d86cb816d6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '978666785b0743bb843de96585f96fb4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 879.887929] env[62965]: DEBUG oslo_vmware.api [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1389900, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.544961} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.888404] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] b59d6023-a047-4b32-bd18-1c7cc05f67d7/b59d6023-a047-4b32-bd18-1c7cc05f67d7.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 879.889020] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 879.889694] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0eadf611-1a71-42aa-a8b3-4c4acad25b9c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.897640] env[62965]: DEBUG oslo_vmware.api [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 879.897640] env[62965]: value = "task-1389905" [ 879.897640] env[62965]: _type = "Task" [ 879.897640] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.906405] env[62965]: DEBUG oslo_vmware.api [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1389905, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.953051] env[62965]: DEBUG nova.network.neutron [req-03a27643-c7fd-4a18-ba55-8c1c77a19ff4 req-90985482-7e18-47b1-9e1f-9a61cbd13fac service nova] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Updated VIF entry in instance network info cache for port 80afc0d2-0d83-4374-a4c0-792eabf66c40. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 879.953051] env[62965]: DEBUG nova.network.neutron [req-03a27643-c7fd-4a18-ba55-8c1c77a19ff4 req-90985482-7e18-47b1-9e1f-9a61cbd13fac service nova] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Updating instance_info_cache with network_info: [{"id": "80afc0d2-0d83-4374-a4c0-792eabf66c40", "address": "fa:16:3e:1c:0d:5e", "network": {"id": "7ef06255-01ec-4620-8f3d-6085875aa8a3", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-118464573-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.223", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1caebff08644e21b3f1c461aa115b8e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap80afc0d2-0d", "ovs_interfaceid": "80afc0d2-0d83-4374-a4c0-792eabf66c40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.975944] env[62965]: DEBUG nova.compute.manager [req-17a6eff0-7b67-41b3-b51c-3b0b31688902 req-7e2b03d7-674f-4e02-941e-e274e902acf6 service nova] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Received event network-vif-plugged-6ffa156f-6a68-424e-b324-660892b9ad22 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 879.976695] env[62965]: DEBUG oslo_concurrency.lockutils [req-17a6eff0-7b67-41b3-b51c-3b0b31688902 req-7e2b03d7-674f-4e02-941e-e274e902acf6 service nova] Acquiring lock "a0e9cc1a-bc95-4a44-99c5-aabf85a373f4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.977085] env[62965]: DEBUG oslo_concurrency.lockutils [req-17a6eff0-7b67-41b3-b51c-3b0b31688902 req-7e2b03d7-674f-4e02-941e-e274e902acf6 service nova] Lock "a0e9cc1a-bc95-4a44-99c5-aabf85a373f4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.977379] env[62965]: DEBUG oslo_concurrency.lockutils [req-17a6eff0-7b67-41b3-b51c-3b0b31688902 req-7e2b03d7-674f-4e02-941e-e274e902acf6 service nova] Lock "a0e9cc1a-bc95-4a44-99c5-aabf85a373f4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.977705] env[62965]: DEBUG nova.compute.manager [req-17a6eff0-7b67-41b3-b51c-3b0b31688902 req-7e2b03d7-674f-4e02-941e-e274e902acf6 service nova] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] No waiting events found dispatching network-vif-plugged-6ffa156f-6a68-424e-b324-660892b9ad22 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 879.978074] env[62965]: WARNING nova.compute.manager [req-17a6eff0-7b67-41b3-b51c-3b0b31688902 req-7e2b03d7-674f-4e02-941e-e274e902acf6 service nova] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Received unexpected event network-vif-plugged-6ffa156f-6a68-424e-b324-660892b9ad22 for instance with vm_state building and task_state spawning. [ 880.091242] env[62965]: DEBUG oslo_vmware.api [None req-320eb21f-761c-4700-80cc-9f87a0353092 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1389901, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.210637] env[62965]: DEBUG nova.network.neutron [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Successfully created port: 4ab65d3b-4168-447e-b02a-0205186fa185 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 880.313962] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquiring lock "refresh_cache-a0e9cc1a-bc95-4a44-99c5-aabf85a373f4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.314602] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquired lock "refresh_cache-a0e9cc1a-bc95-4a44-99c5-aabf85a373f4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.314602] env[62965]: DEBUG nova.network.neutron [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 880.320883] env[62965]: DEBUG nova.compute.manager [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 880.374488] env[62965]: DEBUG oslo_vmware.api [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]520ed17e-91b9-afbb-3cbf-dae27a1b10b0, 'name': SearchDatastore_Task, 'duration_secs': 0.010299} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.374912] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.375414] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 880.375726] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.375984] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.376178] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 880.378853] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9e017d31-1322-4b5b-a4b9-21317f39bf8e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.388364] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 880.389539] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 880.389539] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-477cefc0-2389-4869-830b-272f7e9dfb5c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.397969] env[62965]: DEBUG oslo_vmware.api [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 880.397969] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]525724f8-a297-17fe-f61f-507f09d607b3" [ 880.397969] env[62965]: _type = "Task" [ 880.397969] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.421128] env[62965]: DEBUG oslo_vmware.api [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]525724f8-a297-17fe-f61f-507f09d607b3, 'name': SearchDatastore_Task, 'duration_secs': 0.009933} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.421453] env[62965]: DEBUG oslo_vmware.api [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1389905, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087904} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.422506] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 880.422825] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba3777d4-b61e-4c9d-a084-f10e16bb5146 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.425610] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a28334b8-ec4d-4798-a99f-a2dd1235e965 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.432016] env[62965]: DEBUG oslo_vmware.api [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 880.432016] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52027031-cb6b-b0b6-5d36-28d1a7663c79" [ 880.432016] env[62965]: _type = "Task" [ 880.432016] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.455387] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] b59d6023-a047-4b32-bd18-1c7cc05f67d7/b59d6023-a047-4b32-bd18-1c7cc05f67d7.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 880.461821] env[62965]: DEBUG oslo_concurrency.lockutils [req-03a27643-c7fd-4a18-ba55-8c1c77a19ff4 req-90985482-7e18-47b1-9e1f-9a61cbd13fac service nova] Releasing lock "refresh_cache-924c5cdc-0450-4ced-b920-50e2d5060fd2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.462093] env[62965]: DEBUG nova.compute.manager [req-03a27643-c7fd-4a18-ba55-8c1c77a19ff4 req-90985482-7e18-47b1-9e1f-9a61cbd13fac service nova] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Received event network-vif-plugged-7faa0a61-99a2-4ae4-b60a-ef2eba7aef31 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 880.462290] env[62965]: DEBUG oslo_concurrency.lockutils [req-03a27643-c7fd-4a18-ba55-8c1c77a19ff4 req-90985482-7e18-47b1-9e1f-9a61cbd13fac service nova] Acquiring lock "bd36601b-5a70-4a29-8ef1-d1e925f41de7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.462493] env[62965]: DEBUG oslo_concurrency.lockutils [req-03a27643-c7fd-4a18-ba55-8c1c77a19ff4 req-90985482-7e18-47b1-9e1f-9a61cbd13fac service nova] Lock "bd36601b-5a70-4a29-8ef1-d1e925f41de7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.462672] env[62965]: DEBUG oslo_concurrency.lockutils [req-03a27643-c7fd-4a18-ba55-8c1c77a19ff4 req-90985482-7e18-47b1-9e1f-9a61cbd13fac service nova] Lock "bd36601b-5a70-4a29-8ef1-d1e925f41de7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.462820] env[62965]: DEBUG nova.compute.manager [req-03a27643-c7fd-4a18-ba55-8c1c77a19ff4 req-90985482-7e18-47b1-9e1f-9a61cbd13fac service nova] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] No waiting events found dispatching network-vif-plugged-7faa0a61-99a2-4ae4-b60a-ef2eba7aef31 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 880.462990] env[62965]: WARNING nova.compute.manager [req-03a27643-c7fd-4a18-ba55-8c1c77a19ff4 req-90985482-7e18-47b1-9e1f-9a61cbd13fac service nova] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Received unexpected event network-vif-plugged-7faa0a61-99a2-4ae4-b60a-ef2eba7aef31 for instance with vm_state building and task_state spawning. [ 880.463203] env[62965]: DEBUG nova.compute.manager [req-03a27643-c7fd-4a18-ba55-8c1c77a19ff4 req-90985482-7e18-47b1-9e1f-9a61cbd13fac service nova] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Received event network-changed-7faa0a61-99a2-4ae4-b60a-ef2eba7aef31 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 880.463362] env[62965]: DEBUG nova.compute.manager [req-03a27643-c7fd-4a18-ba55-8c1c77a19ff4 req-90985482-7e18-47b1-9e1f-9a61cbd13fac service nova] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Refreshing instance network info cache due to event network-changed-7faa0a61-99a2-4ae4-b60a-ef2eba7aef31. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 880.463548] env[62965]: DEBUG oslo_concurrency.lockutils [req-03a27643-c7fd-4a18-ba55-8c1c77a19ff4 req-90985482-7e18-47b1-9e1f-9a61cbd13fac service nova] Acquiring lock "refresh_cache-bd36601b-5a70-4a29-8ef1-d1e925f41de7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.463687] env[62965]: DEBUG oslo_concurrency.lockutils [req-03a27643-c7fd-4a18-ba55-8c1c77a19ff4 req-90985482-7e18-47b1-9e1f-9a61cbd13fac service nova] Acquired lock "refresh_cache-bd36601b-5a70-4a29-8ef1-d1e925f41de7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.463856] env[62965]: DEBUG nova.network.neutron [req-03a27643-c7fd-4a18-ba55-8c1c77a19ff4 req-90985482-7e18-47b1-9e1f-9a61cbd13fac service nova] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Refreshing network info cache for port 7faa0a61-99a2-4ae4-b60a-ef2eba7aef31 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 880.465676] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-de4b8996-4b37-4597-a671-447336497bd1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.492042] env[62965]: DEBUG oslo_vmware.api [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52027031-cb6b-b0b6-5d36-28d1a7663c79, 'name': SearchDatastore_Task, 'duration_secs': 0.009802} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.493490] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.493756] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] bd36601b-5a70-4a29-8ef1-d1e925f41de7/bd36601b-5a70-4a29-8ef1-d1e925f41de7.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 880.494090] env[62965]: DEBUG oslo_vmware.api [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 880.494090] env[62965]: value = "task-1389906" [ 880.494090] env[62965]: _type = "Task" [ 880.494090] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.496567] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-40366389-dfb0-4534-8d11-721b70fbf8b5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.510851] env[62965]: DEBUG oslo_vmware.api [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1389906, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.512258] env[62965]: DEBUG oslo_vmware.api [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 880.512258] env[62965]: value = "task-1389907" [ 880.512258] env[62965]: _type = "Task" [ 880.512258] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.523949] env[62965]: DEBUG oslo_vmware.api [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1389907, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.591519] env[62965]: DEBUG oslo_vmware.api [None req-320eb21f-761c-4700-80cc-9f87a0353092 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1389901, 'name': RemoveSnapshot_Task, 'duration_secs': 1.041007} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.593864] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-320eb21f-761c-4700-80cc-9f87a0353092 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Deleted Snapshot of the VM instance {{(pid=62965) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 880.614037] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dd97693-43bb-4a77-9d81-4b670f87c1f4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.622509] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f254724-8703-4d77-a545-7d7aa9319949 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.661891] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db9cda05-d0c9-4242-aaea-3fb0be8e516c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.671347] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ce71bdb-f8a5-46d9-8850-fd7379b52330 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.688124] env[62965]: DEBUG nova.compute.provider_tree [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 880.722568] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 880.722919] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5d267fd8-51e4-448b-9fdf-474bf88687b3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.730551] env[62965]: DEBUG oslo_vmware.api [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 880.730551] env[62965]: value = "task-1389908" [ 880.730551] env[62965]: _type = "Task" [ 880.730551] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.739660] env[62965]: DEBUG oslo_vmware.api [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1389908, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.857191] env[62965]: DEBUG nova.network.neutron [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 881.016421] env[62965]: DEBUG oslo_vmware.api [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1389906, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.027163] env[62965]: DEBUG oslo_vmware.api [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1389907, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.063084] env[62965]: DEBUG nova.network.neutron [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Updating instance_info_cache with network_info: [{"id": "6ffa156f-6a68-424e-b324-660892b9ad22", "address": "fa:16:3e:ed:d8:a1", "network": {"id": "e9a6ebdd-0479-45af-b947-5d35ae182c87", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-385052062-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89fd5275b2064288b44e79365f6f0271", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ba07329-1d3e-4ba8-8774-d029262318c4", "external-id": "nsx-vlan-transportzone-534", "segmentation_id": 534, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ffa156f-6a", "ovs_interfaceid": "6ffa156f-6a68-424e-b324-660892b9ad22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.099252] env[62965]: WARNING nova.compute.manager [None req-320eb21f-761c-4700-80cc-9f87a0353092 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Image not found during snapshot: nova.exception.ImageNotFound: Image ed859b32-0dd3-4eca-8639-006d78224e79 could not be found. [ 881.195424] env[62965]: DEBUG nova.scheduler.client.report [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 881.243978] env[62965]: DEBUG oslo_vmware.api [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1389908, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.332740] env[62965]: DEBUG nova.compute.manager [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 881.366570] env[62965]: DEBUG nova.virt.hardware [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 881.367056] env[62965]: DEBUG nova.virt.hardware [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 881.367516] env[62965]: DEBUG nova.virt.hardware [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 881.367845] env[62965]: DEBUG nova.virt.hardware [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 881.368124] env[62965]: DEBUG nova.virt.hardware [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 881.368385] env[62965]: DEBUG nova.virt.hardware [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 881.368727] env[62965]: DEBUG nova.virt.hardware [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 881.368921] env[62965]: DEBUG nova.virt.hardware [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 881.369459] env[62965]: DEBUG nova.virt.hardware [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 881.369990] env[62965]: DEBUG nova.virt.hardware [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 881.370377] env[62965]: DEBUG nova.virt.hardware [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 881.373255] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6545f114-06b6-44dc-a095-8c9e80bd3f7a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.389177] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6545b1b-a1b9-4e68-88c6-31ede2046b68 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.395957] env[62965]: DEBUG nova.compute.manager [req-341481af-0a6b-4c3b-a8bf-5f479801d1c5 req-d26dc259-143f-4ee3-8e39-bf51bc36b0c3 service nova] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Received event network-changed-65a91c43-8ff0-4e66-8f18-2de456875968 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 881.396190] env[62965]: DEBUG nova.compute.manager [req-341481af-0a6b-4c3b-a8bf-5f479801d1c5 req-d26dc259-143f-4ee3-8e39-bf51bc36b0c3 service nova] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Refreshing instance network info cache due to event network-changed-65a91c43-8ff0-4e66-8f18-2de456875968. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 881.396412] env[62965]: DEBUG oslo_concurrency.lockutils [req-341481af-0a6b-4c3b-a8bf-5f479801d1c5 req-d26dc259-143f-4ee3-8e39-bf51bc36b0c3 service nova] Acquiring lock "refresh_cache-6def4a38-d3b8-4724-97a1-1859b239d900" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.396552] env[62965]: DEBUG oslo_concurrency.lockutils [req-341481af-0a6b-4c3b-a8bf-5f479801d1c5 req-d26dc259-143f-4ee3-8e39-bf51bc36b0c3 service nova] Acquired lock "refresh_cache-6def4a38-d3b8-4724-97a1-1859b239d900" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.396900] env[62965]: DEBUG nova.network.neutron [req-341481af-0a6b-4c3b-a8bf-5f479801d1c5 req-d26dc259-143f-4ee3-8e39-bf51bc36b0c3 service nova] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Refreshing network info cache for port 65a91c43-8ff0-4e66-8f18-2de456875968 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 881.493493] env[62965]: DEBUG oslo_concurrency.lockutils [None req-331e5ade-da3c-4f0a-9a4b-a82bb76a893e tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Acquiring lock "b0d8f38a-86d3-466a-9b83-35145071dc21" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.493761] env[62965]: DEBUG oslo_concurrency.lockutils [None req-331e5ade-da3c-4f0a-9a4b-a82bb76a893e tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Lock "b0d8f38a-86d3-466a-9b83-35145071dc21" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.493972] env[62965]: DEBUG oslo_concurrency.lockutils [None req-331e5ade-da3c-4f0a-9a4b-a82bb76a893e tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Acquiring lock "b0d8f38a-86d3-466a-9b83-35145071dc21-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.494187] env[62965]: DEBUG oslo_concurrency.lockutils [None req-331e5ade-da3c-4f0a-9a4b-a82bb76a893e tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Lock "b0d8f38a-86d3-466a-9b83-35145071dc21-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.494371] env[62965]: DEBUG oslo_concurrency.lockutils [None req-331e5ade-da3c-4f0a-9a4b-a82bb76a893e tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Lock "b0d8f38a-86d3-466a-9b83-35145071dc21-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.498207] env[62965]: INFO nova.compute.manager [None req-331e5ade-da3c-4f0a-9a4b-a82bb76a893e tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Terminating instance [ 881.512022] env[62965]: DEBUG oslo_vmware.api [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1389906, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.526542] env[62965]: DEBUG oslo_vmware.api [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1389907, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.762333} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.527469] env[62965]: DEBUG nova.network.neutron [req-03a27643-c7fd-4a18-ba55-8c1c77a19ff4 req-90985482-7e18-47b1-9e1f-9a61cbd13fac service nova] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Updated VIF entry in instance network info cache for port 7faa0a61-99a2-4ae4-b60a-ef2eba7aef31. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 881.527829] env[62965]: DEBUG nova.network.neutron [req-03a27643-c7fd-4a18-ba55-8c1c77a19ff4 req-90985482-7e18-47b1-9e1f-9a61cbd13fac service nova] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Updating instance_info_cache with network_info: [{"id": "7faa0a61-99a2-4ae4-b60a-ef2eba7aef31", "address": "fa:16:3e:ec:07:ba", "network": {"id": "a0a5e605-d580-4b7b-b3c0-5c7395bd5a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1679618017-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ed71548c085499981fa1b102d660368", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7faa0a61-99", "ovs_interfaceid": "7faa0a61-99a2-4ae4-b60a-ef2eba7aef31", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.529042] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] bd36601b-5a70-4a29-8ef1-d1e925f41de7/bd36601b-5a70-4a29-8ef1-d1e925f41de7.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 881.529291] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 881.529788] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-16759cff-ce31-4ff8-9859-af4c457344c8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.538489] env[62965]: DEBUG oslo_vmware.api [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 881.538489] env[62965]: value = "task-1389909" [ 881.538489] env[62965]: _type = "Task" [ 881.538489] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.547117] env[62965]: DEBUG oslo_vmware.api [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1389909, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.566056] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Releasing lock "refresh_cache-a0e9cc1a-bc95-4a44-99c5-aabf85a373f4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.569016] env[62965]: DEBUG nova.compute.manager [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Instance network_info: |[{"id": "6ffa156f-6a68-424e-b324-660892b9ad22", "address": "fa:16:3e:ed:d8:a1", "network": {"id": "e9a6ebdd-0479-45af-b947-5d35ae182c87", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-385052062-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89fd5275b2064288b44e79365f6f0271", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ba07329-1d3e-4ba8-8774-d029262318c4", "external-id": "nsx-vlan-transportzone-534", "segmentation_id": 534, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ffa156f-6a", "ovs_interfaceid": "6ffa156f-6a68-424e-b324-660892b9ad22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 881.569396] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ed:d8:a1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5ba07329-1d3e-4ba8-8774-d029262318c4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6ffa156f-6a68-424e-b324-660892b9ad22', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 881.576132] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Creating folder: Project (89fd5275b2064288b44e79365f6f0271). Parent ref: group-v294931. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 881.576568] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-79f3f534-a6a7-4259-a558-d5472a3b2631 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.588224] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Created folder: Project (89fd5275b2064288b44e79365f6f0271) in parent group-v294931. [ 881.588224] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Creating folder: Instances. Parent ref: group-v294988. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 881.588224] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cf6069d1-f020-43bb-8e8f-6fe110a06864 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.596911] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Created folder: Instances in parent group-v294988. [ 881.597151] env[62965]: DEBUG oslo.service.loopingcall [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 881.597441] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 881.597759] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-664f5922-7a64-4ad5-a202-618c23b0f896 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.618884] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 881.618884] env[62965]: value = "task-1389912" [ 881.618884] env[62965]: _type = "Task" [ 881.618884] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.627114] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389912, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.709167] env[62965]: DEBUG oslo_concurrency.lockutils [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.417s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.709167] env[62965]: DEBUG nova.compute.manager [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 881.712062] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 9.471s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.742023] env[62965]: DEBUG oslo_vmware.api [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1389908, 'name': PowerOffVM_Task, 'duration_secs': 0.615711} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.743034] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 881.743034] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 881.743390] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87464cdc-1289-427c-a1df-eafbb7556333 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.751932] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 881.751932] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4721f1b5-9423-4092-ab62-0f0fddd375c4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.929938] env[62965]: DEBUG nova.network.neutron [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Successfully updated port: 4ab65d3b-4168-447e-b02a-0205186fa185 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 882.007375] env[62965]: DEBUG nova.compute.manager [None req-331e5ade-da3c-4f0a-9a4b-a82bb76a893e tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 882.007751] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-331e5ade-da3c-4f0a-9a4b-a82bb76a893e tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 882.009202] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4b6369c-9a7b-4d49-9532-9ea1d73e8999 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.014787] env[62965]: DEBUG nova.compute.manager [req-4c333450-8320-458f-8488-6aca618bf854 req-157e4397-7621-4f0d-ba1b-8be35a6cd5c2 service nova] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Received event network-changed-6ffa156f-6a68-424e-b324-660892b9ad22 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 882.014787] env[62965]: DEBUG nova.compute.manager [req-4c333450-8320-458f-8488-6aca618bf854 req-157e4397-7621-4f0d-ba1b-8be35a6cd5c2 service nova] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Refreshing instance network info cache due to event network-changed-6ffa156f-6a68-424e-b324-660892b9ad22. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 882.015203] env[62965]: DEBUG oslo_concurrency.lockutils [req-4c333450-8320-458f-8488-6aca618bf854 req-157e4397-7621-4f0d-ba1b-8be35a6cd5c2 service nova] Acquiring lock "refresh_cache-a0e9cc1a-bc95-4a44-99c5-aabf85a373f4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.015203] env[62965]: DEBUG oslo_concurrency.lockutils [req-4c333450-8320-458f-8488-6aca618bf854 req-157e4397-7621-4f0d-ba1b-8be35a6cd5c2 service nova] Acquired lock "refresh_cache-a0e9cc1a-bc95-4a44-99c5-aabf85a373f4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.015396] env[62965]: DEBUG nova.network.neutron [req-4c333450-8320-458f-8488-6aca618bf854 req-157e4397-7621-4f0d-ba1b-8be35a6cd5c2 service nova] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Refreshing network info cache for port 6ffa156f-6a68-424e-b324-660892b9ad22 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 882.024660] env[62965]: DEBUG oslo_vmware.api [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1389906, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.031656] env[62965]: DEBUG oslo_concurrency.lockutils [req-03a27643-c7fd-4a18-ba55-8c1c77a19ff4 req-90985482-7e18-47b1-9e1f-9a61cbd13fac service nova] Releasing lock "refresh_cache-bd36601b-5a70-4a29-8ef1-d1e925f41de7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.032338] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-331e5ade-da3c-4f0a-9a4b-a82bb76a893e tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 882.033234] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2096c4f0-e39a-47f3-b5de-8caee4c63398 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.039588] env[62965]: DEBUG oslo_vmware.api [None req-331e5ade-da3c-4f0a-9a4b-a82bb76a893e tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 882.039588] env[62965]: value = "task-1389914" [ 882.039588] env[62965]: _type = "Task" [ 882.039588] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.054213] env[62965]: DEBUG oslo_vmware.api [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1389909, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.0738} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.057284] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 882.057661] env[62965]: DEBUG oslo_vmware.api [None req-331e5ade-da3c-4f0a-9a4b-a82bb76a893e tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1389914, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.061585] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4260102-74a3-4aef-9d35-1a14fd11616f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.084300] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] bd36601b-5a70-4a29-8ef1-d1e925f41de7/bd36601b-5a70-4a29-8ef1-d1e925f41de7.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 882.086824] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2f129418-34a6-4e0f-9529-aded6f68aa41 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.107366] env[62965]: DEBUG oslo_vmware.api [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 882.107366] env[62965]: value = "task-1389915" [ 882.107366] env[62965]: _type = "Task" [ 882.107366] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.115301] env[62965]: DEBUG oslo_vmware.api [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1389915, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.129556] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389912, 'name': CreateVM_Task, 'duration_secs': 0.437584} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.129763] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 882.130629] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.130821] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.131136] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 882.131448] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a751696a-9a49-44c6-8bd2-e5649b20e56a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.136061] env[62965]: DEBUG oslo_vmware.api [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 882.136061] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]520e4f65-a60d-9065-fff5-41e5cbf003ad" [ 882.136061] env[62965]: _type = "Task" [ 882.136061] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.144415] env[62965]: DEBUG oslo_vmware.api [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]520e4f65-a60d-9065-fff5-41e5cbf003ad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.215565] env[62965]: DEBUG nova.compute.utils [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 882.218239] env[62965]: DEBUG nova.network.neutron [req-341481af-0a6b-4c3b-a8bf-5f479801d1c5 req-d26dc259-143f-4ee3-8e39-bf51bc36b0c3 service nova] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Updated VIF entry in instance network info cache for port 65a91c43-8ff0-4e66-8f18-2de456875968. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 882.218239] env[62965]: DEBUG nova.network.neutron [req-341481af-0a6b-4c3b-a8bf-5f479801d1c5 req-d26dc259-143f-4ee3-8e39-bf51bc36b0c3 service nova] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Updating instance_info_cache with network_info: [{"id": "65a91c43-8ff0-4e66-8f18-2de456875968", "address": "fa:16:3e:9c:2c:4d", "network": {"id": "98b821d6-efa4-46c8-81e6-a703bd34cb6a", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1649479803-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.254", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a539017d4b84f39a71b6f4098395fbe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c979f78-8597-41f8-b1de-995014032689", "external-id": "nsx-vlan-transportzone-477", "segmentation_id": 477, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65a91c43-8f", "ovs_interfaceid": "65a91c43-8ff0-4e66-8f18-2de456875968", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.219487] env[62965]: DEBUG nova.compute.manager [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 882.219487] env[62965]: DEBUG nova.network.neutron [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 882.262319] env[62965]: DEBUG nova.policy [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '182d38ca18c64bed8f3ecc3f95229756', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1952dedf340f4b7ba0d57af6bba9a749', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 882.431936] env[62965]: DEBUG oslo_concurrency.lockutils [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "refresh_cache-4fc543a1-34ce-40bb-9e80-72ab9098eb38" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.432330] env[62965]: DEBUG oslo_concurrency.lockutils [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquired lock "refresh_cache-4fc543a1-34ce-40bb-9e80-72ab9098eb38" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.432330] env[62965]: DEBUG nova.network.neutron [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 882.510979] env[62965]: DEBUG oslo_vmware.api [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1389906, 'name': ReconfigVM_Task, 'duration_secs': 1.880401} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.511638] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Reconfigured VM instance instance-00000048 to attach disk [datastore1] b59d6023-a047-4b32-bd18-1c7cc05f67d7/b59d6023-a047-4b32-bd18-1c7cc05f67d7.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 882.512458] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d826fb01-7898-451f-9f74-e6860e361d64 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.519876] env[62965]: DEBUG oslo_vmware.api [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 882.519876] env[62965]: value = "task-1389916" [ 882.519876] env[62965]: _type = "Task" [ 882.519876] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.532923] env[62965]: DEBUG oslo_vmware.api [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1389916, 'name': Rename_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.561027] env[62965]: DEBUG oslo_vmware.api [None req-331e5ade-da3c-4f0a-9a4b-a82bb76a893e tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1389914, 'name': PowerOffVM_Task, 'duration_secs': 0.485831} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.561027] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-331e5ade-da3c-4f0a-9a4b-a82bb76a893e tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 882.561027] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-331e5ade-da3c-4f0a-9a4b-a82bb76a893e tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 882.561027] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1b3a9c70-4ee5-4870-b781-aed927c68c5f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.611437] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Acquiring lock "fa511b61-887e-46a1-aaf6-586d716c0fb5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.611437] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Lock "fa511b61-887e-46a1-aaf6-586d716c0fb5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.621137] env[62965]: DEBUG oslo_vmware.api [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1389915, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.649911] env[62965]: DEBUG oslo_vmware.api [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]520e4f65-a60d-9065-fff5-41e5cbf003ad, 'name': SearchDatastore_Task, 'duration_secs': 0.008982} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.650337] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.650696] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 882.650945] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.651115] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.651303] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 882.651590] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-aeb2eefd-2e87-4fac-98d3-76f71b349fc4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.672830] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 882.673079] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 882.674092] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-baf50899-166a-44d4-81f1-7716f0fdeaa4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.681070] env[62965]: DEBUG oslo_vmware.api [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 882.681070] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52c1e555-8ace-53a4-ce03-fb1aa4b2d8e0" [ 882.681070] env[62965]: _type = "Task" [ 882.681070] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.692028] env[62965]: DEBUG oslo_vmware.api [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52c1e555-8ace-53a4-ce03-fb1aa4b2d8e0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.727805] env[62965]: DEBUG nova.compute.manager [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 882.730813] env[62965]: DEBUG oslo_concurrency.lockutils [req-341481af-0a6b-4c3b-a8bf-5f479801d1c5 req-d26dc259-143f-4ee3-8e39-bf51bc36b0c3 service nova] Releasing lock "refresh_cache-6def4a38-d3b8-4724-97a1-1859b239d900" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.759315] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 64cf898c-2c97-4daa-80c8-0a1c252cbcf2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 882.759493] env[62965]: WARNING nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 882.759643] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance c2448b76-1553-4b68-a731-0bb0967f4c1d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 882.759740] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance d4d9b202-3aa7-4b66-949d-4a2eb34b3adf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 882.759855] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance b0d8f38a-86d3-466a-9b83-35145071dc21 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 882.759967] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 924c5cdc-0450-4ced-b920-50e2d5060fd2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 882.760109] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 8c478341-9b50-45de-9ba3-321e802ed1f0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 882.760209] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 6def4a38-d3b8-4724-97a1-1859b239d900 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 882.760324] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance b59d6023-a047-4b32-bd18-1c7cc05f67d7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 882.760437] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance bd36601b-5a70-4a29-8ef1-d1e925f41de7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 882.760590] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance a0e9cc1a-bc95-4a44-99c5-aabf85a373f4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 882.760656] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 4fc543a1-34ce-40bb-9e80-72ab9098eb38 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 882.760752] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance d038608c-9264-4fd5-b9ba-0653c98ec0e4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 882.806918] env[62965]: DEBUG nova.network.neutron [req-4c333450-8320-458f-8488-6aca618bf854 req-157e4397-7621-4f0d-ba1b-8be35a6cd5c2 service nova] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Updated VIF entry in instance network info cache for port 6ffa156f-6a68-424e-b324-660892b9ad22. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 882.806918] env[62965]: DEBUG nova.network.neutron [req-4c333450-8320-458f-8488-6aca618bf854 req-157e4397-7621-4f0d-ba1b-8be35a6cd5c2 service nova] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Updating instance_info_cache with network_info: [{"id": "6ffa156f-6a68-424e-b324-660892b9ad22", "address": "fa:16:3e:ed:d8:a1", "network": {"id": "e9a6ebdd-0479-45af-b947-5d35ae182c87", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-385052062-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89fd5275b2064288b44e79365f6f0271", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ba07329-1d3e-4ba8-8774-d029262318c4", "external-id": "nsx-vlan-transportzone-534", "segmentation_id": 534, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ffa156f-6a", "ovs_interfaceid": "6ffa156f-6a68-424e-b324-660892b9ad22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.967164] env[62965]: DEBUG nova.network.neutron [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 883.031024] env[62965]: DEBUG oslo_vmware.api [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1389916, 'name': Rename_Task, 'duration_secs': 0.171773} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.031024] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 883.031024] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-587747c2-6ca9-4d2f-9b02-f28a69314ca0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.038239] env[62965]: DEBUG oslo_vmware.api [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 883.038239] env[62965]: value = "task-1389918" [ 883.038239] env[62965]: _type = "Task" [ 883.038239] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.049368] env[62965]: DEBUG oslo_vmware.api [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1389918, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.115413] env[62965]: DEBUG nova.compute.manager [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 883.123640] env[62965]: DEBUG oslo_vmware.api [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1389915, 'name': ReconfigVM_Task, 'duration_secs': 0.567305} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.124582] env[62965]: DEBUG nova.network.neutron [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Updating instance_info_cache with network_info: [{"id": "4ab65d3b-4168-447e-b02a-0205186fa185", "address": "fa:16:3e:87:d4:98", "network": {"id": "62052700-1781-4933-93c8-fa1aafdb0fd2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1931798228-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "978666785b0743bb843de96585f96fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ab65d3b-41", "ovs_interfaceid": "4ab65d3b-4168-447e-b02a-0205186fa185", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.126159] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Reconfigured VM instance instance-00000049 to attach disk [datastore1] bd36601b-5a70-4a29-8ef1-d1e925f41de7/bd36601b-5a70-4a29-8ef1-d1e925f41de7.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 883.126530] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7e83be82-04e1-4fcb-807b-926790ce81f2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.309449] env[62965]: DEBUG oslo_vmware.api [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 886.309449] env[62965]: value = "task-1389919" [ 886.309449] env[62965]: _type = "Task" [ 886.309449] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.309449] env[62965]: DEBUG oslo_vmware.api [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1389919, 'name': Rename_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.309449] env[62965]: DEBUG oslo_vmware.api [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52c1e555-8ace-53a4-ce03-fb1aa4b2d8e0, 'name': SearchDatastore_Task, 'duration_secs': 0.01055} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.309449] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-836c8f15-13b8-4dbf-8045-75760a3f192d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.309449] env[62965]: DEBUG oslo_vmware.api [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 886.309449] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52da9387-56ea-0a11-049f-20c04935a003" [ 886.309449] env[62965]: _type = "Task" [ 886.309449] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.333551] env[62965]: DEBUG oslo_vmware.api [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52da9387-56ea-0a11-049f-20c04935a003, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.333551] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance fa511b61-887e-46a1-aaf6-586d716c0fb5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 886.333551] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Total usable vcpus: 48, total allocated vcpus: 12 {{(pid=62965) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 886.333551] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2816MB phys_disk=200GB used_disk=12GB total_vcpus=48 used_vcpus=12 pci_stats=[] {{(pid=62965) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 886.333551] env[62965]: DEBUG oslo_concurrency.lockutils [req-4c333450-8320-458f-8488-6aca618bf854 req-157e4397-7621-4f0d-ba1b-8be35a6cd5c2 service nova] Releasing lock "refresh_cache-a0e9cc1a-bc95-4a44-99c5-aabf85a373f4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.335465] env[62965]: DEBUG nova.compute.manager [req-4c333450-8320-458f-8488-6aca618bf854 req-157e4397-7621-4f0d-ba1b-8be35a6cd5c2 service nova] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Received event network-vif-plugged-4ab65d3b-4168-447e-b02a-0205186fa185 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 886.335465] env[62965]: DEBUG oslo_concurrency.lockutils [req-4c333450-8320-458f-8488-6aca618bf854 req-157e4397-7621-4f0d-ba1b-8be35a6cd5c2 service nova] Acquiring lock "4fc543a1-34ce-40bb-9e80-72ab9098eb38-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.335465] env[62965]: DEBUG oslo_concurrency.lockutils [req-4c333450-8320-458f-8488-6aca618bf854 req-157e4397-7621-4f0d-ba1b-8be35a6cd5c2 service nova] Lock "4fc543a1-34ce-40bb-9e80-72ab9098eb38-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.335465] env[62965]: DEBUG oslo_concurrency.lockutils [req-4c333450-8320-458f-8488-6aca618bf854 req-157e4397-7621-4f0d-ba1b-8be35a6cd5c2 service nova] Lock "4fc543a1-34ce-40bb-9e80-72ab9098eb38-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.335465] env[62965]: DEBUG nova.compute.manager [req-4c333450-8320-458f-8488-6aca618bf854 req-157e4397-7621-4f0d-ba1b-8be35a6cd5c2 service nova] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] No waiting events found dispatching network-vif-plugged-4ab65d3b-4168-447e-b02a-0205186fa185 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 886.337469] env[62965]: WARNING nova.compute.manager [req-4c333450-8320-458f-8488-6aca618bf854 req-157e4397-7621-4f0d-ba1b-8be35a6cd5c2 service nova] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Received unexpected event network-vif-plugged-4ab65d3b-4168-447e-b02a-0205186fa185 for instance with vm_state building and task_state spawning. [ 886.337469] env[62965]: DEBUG nova.compute.manager [req-4c333450-8320-458f-8488-6aca618bf854 req-157e4397-7621-4f0d-ba1b-8be35a6cd5c2 service nova] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Received event network-changed-4ab65d3b-4168-447e-b02a-0205186fa185 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 886.337469] env[62965]: DEBUG nova.compute.manager [req-4c333450-8320-458f-8488-6aca618bf854 req-157e4397-7621-4f0d-ba1b-8be35a6cd5c2 service nova] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Refreshing instance network info cache due to event network-changed-4ab65d3b-4168-447e-b02a-0205186fa185. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 886.337469] env[62965]: DEBUG oslo_concurrency.lockutils [req-4c333450-8320-458f-8488-6aca618bf854 req-157e4397-7621-4f0d-ba1b-8be35a6cd5c2 service nova] Acquiring lock "refresh_cache-4fc543a1-34ce-40bb-9e80-72ab9098eb38" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.337469] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15670e19-38a8-4703-8dfe-ad3bb10b1c52 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.337469] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f500c3d-5827-423b-a7b8-45007fb56ed2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.343853] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b59d445e-cc67-49ab-beb1-36f46916b5da {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.343853] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26de87e4-66a9-40dc-af9f-1873b4b3a751 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.343853] env[62965]: DEBUG nova.compute.provider_tree [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 886.343853] env[62965]: DEBUG oslo_vmware.api [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1389918, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.343853] env[62965]: DEBUG oslo_vmware.api [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1389919, 'name': Rename_Task, 'duration_secs': 0.197971} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.344409] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 886.344409] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9a11fccb-17ba-4256-b9fb-3a6d6f84d100 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.344409] env[62965]: DEBUG oslo_vmware.api [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 886.344409] env[62965]: value = "task-1389920" [ 886.344409] env[62965]: _type = "Task" [ 886.344409] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.344409] env[62965]: DEBUG oslo_vmware.api [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1389920, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.344409] env[62965]: DEBUG oslo_vmware.api [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52da9387-56ea-0a11-049f-20c04935a003, 'name': SearchDatastore_Task, 'duration_secs': 0.010144} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.345052] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.345052] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] a0e9cc1a-bc95-4a44-99c5-aabf85a373f4/a0e9cc1a-bc95-4a44-99c5-aabf85a373f4.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 886.345052] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-39bf8ffc-ba39-41e9-ab15-94f92c432f8e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.345052] env[62965]: DEBUG oslo_vmware.api [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 886.345052] env[62965]: value = "task-1389921" [ 886.345052] env[62965]: _type = "Task" [ 886.345052] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.345481] env[62965]: DEBUG oslo_vmware.api [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1389921, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.345481] env[62965]: ERROR nova.scheduler.client.report [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [req-6cd23d6c-f82a-4b8a-87ed-d76d83d632ef] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-6cd23d6c-f82a-4b8a-87ed-d76d83d632ef"}]} [ 886.345481] env[62965]: DEBUG oslo_vmware.api [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1389918, 'name': PowerOnVM_Task, 'duration_secs': 0.572316} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.345736] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 886.345736] env[62965]: INFO nova.compute.manager [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Took 9.86 seconds to spawn the instance on the hypervisor. [ 886.345736] env[62965]: DEBUG nova.compute.manager [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 886.345736] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96b962f9-846b-4d19-9519-2b89f8bba8c3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.345736] env[62965]: DEBUG nova.scheduler.client.report [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Refreshing inventories for resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 886.345999] env[62965]: DEBUG nova.scheduler.client.report [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Updating ProviderTree inventory for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 886.345999] env[62965]: DEBUG nova.compute.provider_tree [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 886.345999] env[62965]: DEBUG nova.scheduler.client.report [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Refreshing aggregate associations for resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8, aggregates: None {{(pid=62965) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 886.346334] env[62965]: DEBUG nova.scheduler.client.report [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Refreshing trait associations for resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64 {{(pid=62965) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 886.346334] env[62965]: DEBUG oslo_vmware.api [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1389920, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.346334] env[62965]: DEBUG oslo_vmware.api [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1389921, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.346334] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2867767d-5e03-4e9e-b1d3-53147408623a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.346334] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0427d227-0636-4cd4-9f48-ce3be4823c02 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.346334] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f80606b-a9bd-4a24-a91d-29f08b68137b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.346639] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db2b5af2-b7ef-4479-affb-4b58aedd3c5c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.346639] env[62965]: DEBUG nova.compute.provider_tree [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 886.346639] env[62965]: DEBUG oslo_vmware.api [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1389920, 'name': PowerOnVM_Task, 'duration_secs': 0.785248} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.346639] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 886.346639] env[62965]: INFO nova.compute.manager [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Took 8.14 seconds to spawn the instance on the hypervisor. [ 886.346921] env[62965]: DEBUG nova.compute.manager [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 886.346921] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-558ae818-5aac-41b8-a76b-45c402c3bae6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.346921] env[62965]: DEBUG oslo_vmware.api [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1389921, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.706527} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.346921] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] a0e9cc1a-bc95-4a44-99c5-aabf85a373f4/a0e9cc1a-bc95-4a44-99c5-aabf85a373f4.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 886.346921] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 886.347225] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9abc170b-6f0a-4ef4-ab18-be3b8e4cad9a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.347225] env[62965]: DEBUG oslo_vmware.api [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 886.347225] env[62965]: value = "task-1389922" [ 886.347225] env[62965]: _type = "Task" [ 886.347225] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.347225] env[62965]: DEBUG oslo_vmware.api [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1389922, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.347225] env[62965]: DEBUG oslo_vmware.api [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1389922, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.105607} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.347225] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 886.347777] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf054f2c-5bff-4656-8dff-1682e7d47f47 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.347777] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] a0e9cc1a-bc95-4a44-99c5-aabf85a373f4/a0e9cc1a-bc95-4a44-99c5-aabf85a373f4.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 886.347777] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fe5da6fc-2724-4c45-b020-eb7e934b41c9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.347777] env[62965]: DEBUG oslo_vmware.api [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 886.347777] env[62965]: value = "task-1389923" [ 886.347777] env[62965]: _type = "Task" [ 886.347777] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.347777] env[62965]: DEBUG oslo_vmware.api [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1389923, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.348171] env[62965]: DEBUG oslo_vmware.api [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1389923, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.348171] env[62965]: DEBUG oslo_vmware.api [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1389923, 'name': ReconfigVM_Task, 'duration_secs': 0.60743} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.348171] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Reconfigured VM instance instance-0000004a to attach disk [datastore1] a0e9cc1a-bc95-4a44-99c5-aabf85a373f4/a0e9cc1a-bc95-4a44-99c5-aabf85a373f4.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 886.348171] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6c78b6db-b7c9-4383-ba6c-127d9962cca3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.348171] env[62965]: DEBUG oslo_vmware.api [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 886.348171] env[62965]: value = "task-1389924" [ 886.348171] env[62965]: _type = "Task" [ 886.348171] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.348527] env[62965]: DEBUG nova.network.neutron [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Successfully created port: b1718061-edf1-4077-b57f-0e54aea365a6 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 886.348527] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-331e5ade-da3c-4f0a-9a4b-a82bb76a893e tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 886.348527] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-331e5ade-da3c-4f0a-9a4b-a82bb76a893e tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Deleting contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 886.348527] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-331e5ade-da3c-4f0a-9a4b-a82bb76a893e tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Deleting the datastore file [datastore2] b0d8f38a-86d3-466a-9b83-35145071dc21 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 886.348527] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 886.348753] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Deleting contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 886.348753] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Deleting the datastore file [datastore2] 8c478341-9b50-45de-9ba3-321e802ed1f0 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 886.348753] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-45a8232f-b8c5-465c-b065-2a55a33c62fa {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.348753] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8b091686-3afb-4e46-a079-c7cfa84864a7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.348753] env[62965]: DEBUG oslo_vmware.api [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1389924, 'name': Rename_Task} progress is 10%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.348753] env[62965]: DEBUG oslo_vmware.api [None req-331e5ade-da3c-4f0a-9a4b-a82bb76a893e tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 886.348753] env[62965]: value = "task-1389925" [ 886.348753] env[62965]: _type = "Task" [ 886.351419] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.351419] env[62965]: DEBUG oslo_vmware.api [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 886.351419] env[62965]: value = "task-1389926" [ 886.351419] env[62965]: _type = "Task" [ 886.351419] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.369129] env[62965]: DEBUG oslo_vmware.api [None req-331e5ade-da3c-4f0a-9a4b-a82bb76a893e tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1389925, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.370869] env[62965]: DEBUG oslo_vmware.api [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1389926, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.395035] env[62965]: DEBUG nova.scheduler.client.report [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Updated inventory for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with generation 83 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 886.395035] env[62965]: DEBUG nova.compute.provider_tree [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Updating resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 generation from 83 to 84 during operation: update_inventory {{(pid=62965) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 886.395221] env[62965]: DEBUG nova.compute.provider_tree [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 886.471955] env[62965]: DEBUG oslo_vmware.rw_handles [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52be7c0e-3874-7f2c-a856-0fff94f43cde/disk-0.vmdk. {{(pid=62965) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 886.472716] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9218bf93-11b1-43f8-bbb9-179f8bcceb1c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.481206] env[62965]: DEBUG oslo_vmware.rw_handles [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52be7c0e-3874-7f2c-a856-0fff94f43cde/disk-0.vmdk is in state: ready. {{(pid=62965) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 886.481206] env[62965]: ERROR oslo_vmware.rw_handles [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52be7c0e-3874-7f2c-a856-0fff94f43cde/disk-0.vmdk due to incomplete transfer. [ 886.481418] env[62965]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-42392f6b-a048-4c43-b09a-fe5bcd3551c2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.488490] env[62965]: DEBUG oslo_vmware.rw_handles [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52be7c0e-3874-7f2c-a856-0fff94f43cde/disk-0.vmdk. {{(pid=62965) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 886.488689] env[62965]: DEBUG nova.virt.vmwareapi.images [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Uploaded image 7e436a80-d1ff-4c42-847a-9538961b1a6a to the Glance image server {{(pid=62965) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 886.491662] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Destroying the VM {{(pid=62965) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 886.492059] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-5fd90077-e8d2-435f-92fe-21913039f30f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.498659] env[62965]: DEBUG oslo_vmware.api [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 886.498659] env[62965]: value = "task-1389927" [ 886.498659] env[62965]: _type = "Task" [ 886.498659] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.511998] env[62965]: DEBUG oslo_vmware.api [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389927, 'name': Destroy_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.588539] env[62965]: INFO nova.compute.manager [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Took 30.00 seconds to build instance. [ 886.634533] env[62965]: DEBUG oslo_concurrency.lockutils [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Releasing lock "refresh_cache-4fc543a1-34ce-40bb-9e80-72ab9098eb38" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.634858] env[62965]: DEBUG nova.compute.manager [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Instance network_info: |[{"id": "4ab65d3b-4168-447e-b02a-0205186fa185", "address": "fa:16:3e:87:d4:98", "network": {"id": "62052700-1781-4933-93c8-fa1aafdb0fd2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1931798228-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "978666785b0743bb843de96585f96fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ab65d3b-41", "ovs_interfaceid": "4ab65d3b-4168-447e-b02a-0205186fa185", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 886.635227] env[62965]: DEBUG oslo_concurrency.lockutils [req-4c333450-8320-458f-8488-6aca618bf854 req-157e4397-7621-4f0d-ba1b-8be35a6cd5c2 service nova] Acquired lock "refresh_cache-4fc543a1-34ce-40bb-9e80-72ab9098eb38" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.635416] env[62965]: DEBUG nova.network.neutron [req-4c333450-8320-458f-8488-6aca618bf854 req-157e4397-7621-4f0d-ba1b-8be35a6cd5c2 service nova] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Refreshing network info cache for port 4ab65d3b-4168-447e-b02a-0205186fa185 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 886.636554] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:87:d4:98', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a10c88d7-d13f-44fd-acee-7a734eb5f56a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4ab65d3b-4168-447e-b02a-0205186fa185', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 886.645953] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Creating folder: Project (978666785b0743bb843de96585f96fb4). Parent ref: group-v294931. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 886.648025] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-eb54e836-0cae-44d8-8e44-33411090c1ae {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.655075] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.661682] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Created folder: Project (978666785b0743bb843de96585f96fb4) in parent group-v294931. [ 886.661886] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Creating folder: Instances. Parent ref: group-v294991. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 886.662239] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9486a5e7-2ccd-4632-8369-8b59850bfb3e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.671552] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Created folder: Instances in parent group-v294991. [ 886.671794] env[62965]: DEBUG oslo.service.loopingcall [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 886.672022] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 886.672242] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ebf12c8d-1401-4ea6-bf17-2a1773332aef {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.693848] env[62965]: INFO nova.compute.manager [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Took 22.96 seconds to build instance. [ 886.696886] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 886.696886] env[62965]: value = "task-1389930" [ 886.696886] env[62965]: _type = "Task" [ 886.696886] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.704770] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389930, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.746701] env[62965]: DEBUG nova.compute.manager [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 886.777122] env[62965]: DEBUG nova.virt.hardware [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 886.777426] env[62965]: DEBUG nova.virt.hardware [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 886.777679] env[62965]: DEBUG nova.virt.hardware [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 886.777936] env[62965]: DEBUG nova.virt.hardware [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 886.778112] env[62965]: DEBUG nova.virt.hardware [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 886.778275] env[62965]: DEBUG nova.virt.hardware [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 886.778485] env[62965]: DEBUG nova.virt.hardware [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 886.778648] env[62965]: DEBUG nova.virt.hardware [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 886.778821] env[62965]: DEBUG nova.virt.hardware [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 886.778983] env[62965]: DEBUG nova.virt.hardware [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 886.779177] env[62965]: DEBUG nova.virt.hardware [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 886.780445] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a16f887-90ec-4da6-9b6d-1161780a3286 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.788759] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-502c2978-0d99-4bdb-b69f-383ebc27850b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.821054] env[62965]: DEBUG oslo_vmware.api [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1389924, 'name': Rename_Task, 'duration_secs': 0.198348} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.821308] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 886.821554] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2dc145eb-f9cf-4ef4-86f9-ba6efadfb2d5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.827288] env[62965]: DEBUG oslo_vmware.api [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 886.827288] env[62965]: value = "task-1389931" [ 886.827288] env[62965]: _type = "Task" [ 886.827288] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.834452] env[62965]: DEBUG oslo_vmware.api [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1389931, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.856214] env[62965]: DEBUG oslo_vmware.api [None req-331e5ade-da3c-4f0a-9a4b-a82bb76a893e tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1389925, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.452475} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.859320] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-331e5ade-da3c-4f0a-9a4b-a82bb76a893e tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 886.859490] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-331e5ade-da3c-4f0a-9a4b-a82bb76a893e tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Deleted contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 886.859686] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-331e5ade-da3c-4f0a-9a4b-a82bb76a893e tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 886.859882] env[62965]: INFO nova.compute.manager [None req-331e5ade-da3c-4f0a-9a4b-a82bb76a893e tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Took 4.85 seconds to destroy the instance on the hypervisor. [ 886.860136] env[62965]: DEBUG oslo.service.loopingcall [None req-331e5ade-da3c-4f0a-9a4b-a82bb76a893e tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 886.860346] env[62965]: DEBUG nova.compute.manager [-] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 886.860440] env[62965]: DEBUG nova.network.neutron [-] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 886.866941] env[62965]: DEBUG oslo_vmware.api [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1389926, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.455624} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.867245] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 886.867448] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Deleted contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 886.867647] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 886.900626] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62965) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 886.900845] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 5.189s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.901920] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 13.228s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.902147] env[62965]: DEBUG nova.objects.instance [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62965) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 887.008568] env[62965]: DEBUG oslo_vmware.api [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389927, 'name': Destroy_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.091246] env[62965]: DEBUG oslo_concurrency.lockutils [None req-43424f72-b587-4b10-82a7-a6e446e91656 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "b59d6023-a047-4b32-bd18-1c7cc05f67d7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.016s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.196016] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f078fb7b-616a-4fbf-8a17-8d7a53ca707f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "bd36601b-5a70-4a29-8ef1-d1e925f41de7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.121s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.208861] env[62965]: DEBUG nova.compute.manager [req-9fb9ad1a-1f91-45a0-99e9-25d50a0a51fd req-f79c2d74-c9a5-4407-b8c1-029a3892d901 service nova] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Received event network-vif-deleted-2cf58a6d-2155-49a4-a049-98ef841cc561 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 887.209177] env[62965]: INFO nova.compute.manager [req-9fb9ad1a-1f91-45a0-99e9-25d50a0a51fd req-f79c2d74-c9a5-4407-b8c1-029a3892d901 service nova] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Neutron deleted interface 2cf58a6d-2155-49a4-a049-98ef841cc561; detaching it from the instance and deleting it from the info cache [ 887.209437] env[62965]: DEBUG nova.network.neutron [req-9fb9ad1a-1f91-45a0-99e9-25d50a0a51fd req-f79c2d74-c9a5-4407-b8c1-029a3892d901 service nova] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.217151] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389930, 'name': CreateVM_Task} progress is 99%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.337686] env[62965]: DEBUG oslo_vmware.api [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1389931, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.430406] env[62965]: DEBUG nova.network.neutron [req-4c333450-8320-458f-8488-6aca618bf854 req-157e4397-7621-4f0d-ba1b-8be35a6cd5c2 service nova] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Updated VIF entry in instance network info cache for port 4ab65d3b-4168-447e-b02a-0205186fa185. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 887.430764] env[62965]: DEBUG nova.network.neutron [req-4c333450-8320-458f-8488-6aca618bf854 req-157e4397-7621-4f0d-ba1b-8be35a6cd5c2 service nova] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Updating instance_info_cache with network_info: [{"id": "4ab65d3b-4168-447e-b02a-0205186fa185", "address": "fa:16:3e:87:d4:98", "network": {"id": "62052700-1781-4933-93c8-fa1aafdb0fd2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1931798228-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "978666785b0743bb843de96585f96fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ab65d3b-41", "ovs_interfaceid": "4ab65d3b-4168-447e-b02a-0205186fa185", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.510442] env[62965]: DEBUG oslo_vmware.api [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389927, 'name': Destroy_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.667897] env[62965]: DEBUG nova.network.neutron [-] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.708082] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389930, 'name': CreateVM_Task, 'duration_secs': 0.606333} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.708262] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 887.708940] env[62965]: DEBUG oslo_concurrency.lockutils [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.709138] env[62965]: DEBUG oslo_concurrency.lockutils [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.709479] env[62965]: DEBUG oslo_concurrency.lockutils [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 887.709787] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-708c78fb-8ee9-4859-b600-57dac8e2611f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.714447] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-18275cd1-d371-4577-914a-cc9eb26b1379 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.716501] env[62965]: DEBUG oslo_vmware.api [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 887.716501] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52ab9d06-2b9a-0172-10b3-216e2537a541" [ 887.716501] env[62965]: _type = "Task" [ 887.716501] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.725392] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-600d1fa6-be14-47e7-a86b-c4a91aa74c1e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.739832] env[62965]: DEBUG oslo_vmware.api [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52ab9d06-2b9a-0172-10b3-216e2537a541, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.759918] env[62965]: DEBUG nova.compute.manager [req-9fb9ad1a-1f91-45a0-99e9-25d50a0a51fd req-f79c2d74-c9a5-4407-b8c1-029a3892d901 service nova] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Detach interface failed, port_id=2cf58a6d-2155-49a4-a049-98ef841cc561, reason: Instance b0d8f38a-86d3-466a-9b83-35145071dc21 could not be found. {{(pid=62965) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11294}} [ 887.838697] env[62965]: DEBUG oslo_vmware.api [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1389931, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.911062] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b808940b-6040-4c25-9006-e4d120c9f1d5 tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.008s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.911417] env[62965]: DEBUG oslo_concurrency.lockutils [None req-95029734-558f-483b-a8bc-93b5504885ef tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.011s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.911648] env[62965]: DEBUG oslo_concurrency.lockutils [None req-95029734-558f-483b-a8bc-93b5504885ef tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.915235] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.260s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.916802] env[62965]: INFO nova.compute.claims [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 887.923944] env[62965]: DEBUG nova.virt.hardware [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 887.923944] env[62965]: DEBUG nova.virt.hardware [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 887.923944] env[62965]: DEBUG nova.virt.hardware [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 887.924889] env[62965]: DEBUG nova.virt.hardware [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 887.924889] env[62965]: DEBUG nova.virt.hardware [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 887.924889] env[62965]: DEBUG nova.virt.hardware [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 887.924889] env[62965]: DEBUG nova.virt.hardware [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 887.924889] env[62965]: DEBUG nova.virt.hardware [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 887.925694] env[62965]: DEBUG nova.virt.hardware [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 887.925694] env[62965]: DEBUG nova.virt.hardware [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 887.925694] env[62965]: DEBUG nova.virt.hardware [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 887.925694] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f20a46c1-1bf6-459e-b7c6-2535a2f92b57 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.934277] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88341416-17d0-4bd9-83b0-59034383a021 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.941545] env[62965]: DEBUG oslo_concurrency.lockutils [req-4c333450-8320-458f-8488-6aca618bf854 req-157e4397-7621-4f0d-ba1b-8be35a6cd5c2 service nova] Releasing lock "refresh_cache-4fc543a1-34ce-40bb-9e80-72ab9098eb38" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.955769] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0d:28:72', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '61b8f0db-488e-42d7-bf6c-6c1665cd5616', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c0f50257-3550-4368-8ad1-aa07da3d6e67', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 887.965819] env[62965]: DEBUG oslo.service.loopingcall [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 887.967050] env[62965]: INFO nova.scheduler.client.report [None req-95029734-558f-483b-a8bc-93b5504885ef tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Deleted allocations for instance d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c [ 887.969141] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 887.970076] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5eab4fd5-4f7e-4952-a079-2856cbaf5d60 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.992126] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 887.992126] env[62965]: value = "task-1389932" [ 887.992126] env[62965]: _type = "Task" [ 887.992126] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.001322] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389932, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.009016] env[62965]: DEBUG oslo_vmware.api [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389927, 'name': Destroy_Task, 'duration_secs': 1.39239} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.009285] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Destroyed the VM [ 888.009528] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Deleting Snapshot of the VM instance {{(pid=62965) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 888.009833] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-965e3ae6-3fbd-486c-a3fd-29a5f52bbbc2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.016239] env[62965]: DEBUG oslo_vmware.api [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 888.016239] env[62965]: value = "task-1389933" [ 888.016239] env[62965]: _type = "Task" [ 888.016239] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.024505] env[62965]: DEBUG oslo_vmware.api [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389933, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.174888] env[62965]: INFO nova.compute.manager [-] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Took 1.31 seconds to deallocate network for instance. [ 888.187400] env[62965]: DEBUG nova.compute.manager [req-1440565f-2b37-4f8c-a491-bd0e84820a34 req-cb59c052-3c89-4425-9ea9-d56981400653 service nova] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Received event network-changed-d2d44211-6856-42ed-8f9f-3b1a58b31349 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 888.187603] env[62965]: DEBUG nova.compute.manager [req-1440565f-2b37-4f8c-a491-bd0e84820a34 req-cb59c052-3c89-4425-9ea9-d56981400653 service nova] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Refreshing instance network info cache due to event network-changed-d2d44211-6856-42ed-8f9f-3b1a58b31349. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 888.187890] env[62965]: DEBUG oslo_concurrency.lockutils [req-1440565f-2b37-4f8c-a491-bd0e84820a34 req-cb59c052-3c89-4425-9ea9-d56981400653 service nova] Acquiring lock "refresh_cache-b59d6023-a047-4b32-bd18-1c7cc05f67d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.187982] env[62965]: DEBUG oslo_concurrency.lockutils [req-1440565f-2b37-4f8c-a491-bd0e84820a34 req-cb59c052-3c89-4425-9ea9-d56981400653 service nova] Acquired lock "refresh_cache-b59d6023-a047-4b32-bd18-1c7cc05f67d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.188169] env[62965]: DEBUG nova.network.neutron [req-1440565f-2b37-4f8c-a491-bd0e84820a34 req-cb59c052-3c89-4425-9ea9-d56981400653 service nova] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Refreshing network info cache for port d2d44211-6856-42ed-8f9f-3b1a58b31349 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 888.229220] env[62965]: DEBUG oslo_vmware.api [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52ab9d06-2b9a-0172-10b3-216e2537a541, 'name': SearchDatastore_Task, 'duration_secs': 0.011089} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.229641] env[62965]: DEBUG oslo_concurrency.lockutils [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.230011] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 888.230333] env[62965]: DEBUG oslo_concurrency.lockutils [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.230552] env[62965]: DEBUG oslo_concurrency.lockutils [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.230746] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 888.231068] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-286161b3-a137-47e1-88f2-142dbf98926d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.239966] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 888.240207] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 888.241129] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3511ff2f-7096-478b-a3ff-80ce3fe552ce {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.250282] env[62965]: DEBUG oslo_vmware.api [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 888.250282] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52d5e296-a97c-6e6f-3140-35c21f25e8ec" [ 888.250282] env[62965]: _type = "Task" [ 888.250282] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.257664] env[62965]: DEBUG oslo_vmware.api [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52d5e296-a97c-6e6f-3140-35c21f25e8ec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.343433] env[62965]: DEBUG oslo_vmware.api [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1389931, 'name': PowerOnVM_Task, 'duration_secs': 1.157767} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.343819] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 888.344107] env[62965]: INFO nova.compute.manager [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Took 9.35 seconds to spawn the instance on the hypervisor. [ 888.345123] env[62965]: DEBUG nova.compute.manager [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 888.346363] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68e67e16-1748-4327-8bca-8fca80f96239 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.428775] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquiring lock "f82d86ec-3364-449d-be54-35627082c02b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.429039] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "f82d86ec-3364-449d-be54-35627082c02b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.468531] env[62965]: DEBUG nova.network.neutron [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Successfully updated port: b1718061-edf1-4077-b57f-0e54aea365a6 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 888.491951] env[62965]: DEBUG oslo_concurrency.lockutils [None req-95029734-558f-483b-a8bc-93b5504885ef tempest-ServerShowV257Test-1213832970 tempest-ServerShowV257Test-1213832970-project-member] Lock "d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.946s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.504011] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389932, 'name': CreateVM_Task, 'duration_secs': 0.351353} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.504675] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 888.505348] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.505519] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.505839] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 888.506329] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bcea058c-038d-4c1c-8a6a-9a00644dd0c4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.510945] env[62965]: DEBUG oslo_vmware.api [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 888.510945] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52f71654-56ef-faf1-9668-d776bce37b4d" [ 888.510945] env[62965]: _type = "Task" [ 888.510945] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.519498] env[62965]: DEBUG oslo_vmware.api [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52f71654-56ef-faf1-9668-d776bce37b4d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.527302] env[62965]: DEBUG oslo_vmware.api [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389933, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.682327] env[62965]: DEBUG oslo_concurrency.lockutils [None req-331e5ade-da3c-4f0a-9a4b-a82bb76a893e tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.760131] env[62965]: DEBUG oslo_vmware.api [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52d5e296-a97c-6e6f-3140-35c21f25e8ec, 'name': SearchDatastore_Task, 'duration_secs': 0.011495} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.760991] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e34207a8-d0bc-4cfa-a5e5-7658961bed48 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.767017] env[62965]: DEBUG oslo_vmware.api [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 888.767017] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]527dba4e-53f7-f555-f9ff-69cc2d707a0a" [ 888.767017] env[62965]: _type = "Task" [ 888.767017] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.775569] env[62965]: DEBUG oslo_vmware.api [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]527dba4e-53f7-f555-f9ff-69cc2d707a0a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.873523] env[62965]: INFO nova.compute.manager [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Took 23.92 seconds to build instance. [ 888.934248] env[62965]: DEBUG nova.compute.manager [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 888.978376] env[62965]: DEBUG oslo_concurrency.lockutils [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "refresh_cache-d038608c-9264-4fd5-b9ba-0653c98ec0e4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.978376] env[62965]: DEBUG oslo_concurrency.lockutils [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquired lock "refresh_cache-d038608c-9264-4fd5-b9ba-0653c98ec0e4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.978376] env[62965]: DEBUG nova.network.neutron [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 889.027873] env[62965]: DEBUG oslo_vmware.api [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52f71654-56ef-faf1-9668-d776bce37b4d, 'name': SearchDatastore_Task, 'duration_secs': 0.01045} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.028530] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.028777] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 889.028993] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.034698] env[62965]: DEBUG oslo_vmware.api [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389933, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.103527] env[62965]: DEBUG nova.network.neutron [req-1440565f-2b37-4f8c-a491-bd0e84820a34 req-cb59c052-3c89-4425-9ea9-d56981400653 service nova] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Updated VIF entry in instance network info cache for port d2d44211-6856-42ed-8f9f-3b1a58b31349. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 889.103973] env[62965]: DEBUG nova.network.neutron [req-1440565f-2b37-4f8c-a491-bd0e84820a34 req-cb59c052-3c89-4425-9ea9-d56981400653 service nova] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Updating instance_info_cache with network_info: [{"id": "d2d44211-6856-42ed-8f9f-3b1a58b31349", "address": "fa:16:3e:3b:f0:e9", "network": {"id": "49dd42b0-6397-4fdd-a319-ed920c72a5af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-373172566-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.230", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d91c99b14018469e8998fd1ccc0fab4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2d44211-68", "ovs_interfaceid": "d2d44211-6856-42ed-8f9f-3b1a58b31349", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.209017] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ce1cbfc-0f56-4b9f-aa71-f26d43c1e741 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.218019] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b1b44fb-d7d1-4aba-a1a7-4508909a8939 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.251089] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5778c665-42b4-4b39-899c-b5e16e82e1cc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.259031] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1a7377b-dbed-4d7d-aada-12fc56333110 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.272874] env[62965]: DEBUG nova.compute.provider_tree [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 889.283362] env[62965]: DEBUG oslo_vmware.api [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]527dba4e-53f7-f555-f9ff-69cc2d707a0a, 'name': SearchDatastore_Task, 'duration_secs': 0.020261} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.284272] env[62965]: DEBUG oslo_concurrency.lockutils [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.284492] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 4fc543a1-34ce-40bb-9e80-72ab9098eb38/4fc543a1-34ce-40bb-9e80-72ab9098eb38.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 889.284769] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.284962] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 889.285189] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a6b49165-491b-4e67-9904-6c396bb8ed25 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.287403] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a1f5124e-62c4-4e12-aa8c-c8d9449e97ea {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.293593] env[62965]: DEBUG oslo_vmware.api [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 889.293593] env[62965]: value = "task-1389934" [ 889.293593] env[62965]: _type = "Task" [ 889.293593] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.297864] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 889.298079] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 889.299034] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c931e65-6cbf-4422-a07e-fb0e2c9cf663 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.303959] env[62965]: DEBUG oslo_vmware.api [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1389934, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.306645] env[62965]: DEBUG oslo_vmware.api [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 889.306645] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5207b83a-f23d-721c-c0b6-356f085da7e2" [ 889.306645] env[62965]: _type = "Task" [ 889.306645] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.313835] env[62965]: DEBUG oslo_vmware.api [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5207b83a-f23d-721c-c0b6-356f085da7e2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.379383] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e1b73eb8-2449-4492-ab42-11d7b50537d9 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lock "a0e9cc1a-bc95-4a44-99c5-aabf85a373f4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.457s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.464307] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.519764] env[62965]: DEBUG nova.network.neutron [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 889.530560] env[62965]: DEBUG oslo_vmware.api [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389933, 'name': RemoveSnapshot_Task, 'duration_secs': 1.120735} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.530863] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Deleted Snapshot of the VM instance {{(pid=62965) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 889.531113] env[62965]: INFO nova.compute.manager [None req-933048be-1e3d-40e5-8583-ab2872e8639c tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Took 19.10 seconds to snapshot the instance on the hypervisor. [ 889.608866] env[62965]: DEBUG oslo_concurrency.lockutils [req-1440565f-2b37-4f8c-a491-bd0e84820a34 req-cb59c052-3c89-4425-9ea9-d56981400653 service nova] Releasing lock "refresh_cache-b59d6023-a047-4b32-bd18-1c7cc05f67d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.657364] env[62965]: DEBUG oslo_concurrency.lockutils [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "892b1302-f536-4dcd-9227-a545ae1ac731" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.657726] env[62965]: DEBUG oslo_concurrency.lockutils [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "892b1302-f536-4dcd-9227-a545ae1ac731" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.677485] env[62965]: DEBUG nova.network.neutron [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Updating instance_info_cache with network_info: [{"id": "b1718061-edf1-4077-b57f-0e54aea365a6", "address": "fa:16:3e:1f:27:16", "network": {"id": "a9699d8b-5e76-4666-a236-33ab8f22345e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1329540850-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1952dedf340f4b7ba0d57af6bba9a749", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a316376e-2ef0-4b1e-b40c-10321ebd7e1a", "external-id": "nsx-vlan-transportzone-942", "segmentation_id": 942, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1718061-ed", "ovs_interfaceid": "b1718061-edf1-4077-b57f-0e54aea365a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.800442] env[62965]: ERROR nova.scheduler.client.report [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] [req-e1c9bfb7-0a31-4996-adf3-ee5b8ac6c5f0] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-e1c9bfb7-0a31-4996-adf3-ee5b8ac6c5f0"}]} [ 889.808550] env[62965]: DEBUG oslo_vmware.api [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1389934, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.819294] env[62965]: DEBUG oslo_vmware.api [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5207b83a-f23d-721c-c0b6-356f085da7e2, 'name': SearchDatastore_Task, 'duration_secs': 0.037178} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.820425] env[62965]: DEBUG nova.scheduler.client.report [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Refreshing inventories for resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 889.823167] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1268353-b0d2-484b-9855-605aa6c722cc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.829207] env[62965]: DEBUG oslo_vmware.api [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 889.829207] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5294853c-60e3-8804-6670-ba6daa616caa" [ 889.829207] env[62965]: _type = "Task" [ 889.829207] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.839360] env[62965]: DEBUG oslo_vmware.api [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5294853c-60e3-8804-6670-ba6daa616caa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.840484] env[62965]: DEBUG nova.scheduler.client.report [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Updating ProviderTree inventory for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 889.840693] env[62965]: DEBUG nova.compute.provider_tree [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 889.853257] env[62965]: DEBUG nova.scheduler.client.report [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Refreshing aggregate associations for resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8, aggregates: None {{(pid=62965) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 889.872057] env[62965]: DEBUG nova.scheduler.client.report [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Refreshing trait associations for resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64 {{(pid=62965) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 890.042793] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquiring lock "eed55d8e-21dc-47d4-a689-415ef6a3891d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.042964] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lock "eed55d8e-21dc-47d4-a689-415ef6a3891d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.092326] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e0a155a-bfd4-49da-83ca-6940fe28258d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.101983] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e78629d-8aac-4be9-a965-d86d0950dfc7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.132122] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f9fe50f-9868-4786-abcc-86f50f78a689 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.139159] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d077b992-4a65-4e42-8380-a70c64281999 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.152321] env[62965]: DEBUG nova.compute.provider_tree [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 890.159551] env[62965]: DEBUG nova.compute.manager [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 890.180557] env[62965]: DEBUG oslo_concurrency.lockutils [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Releasing lock "refresh_cache-d038608c-9264-4fd5-b9ba-0653c98ec0e4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.180557] env[62965]: DEBUG nova.compute.manager [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Instance network_info: |[{"id": "b1718061-edf1-4077-b57f-0e54aea365a6", "address": "fa:16:3e:1f:27:16", "network": {"id": "a9699d8b-5e76-4666-a236-33ab8f22345e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1329540850-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1952dedf340f4b7ba0d57af6bba9a749", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a316376e-2ef0-4b1e-b40c-10321ebd7e1a", "external-id": "nsx-vlan-transportzone-942", "segmentation_id": 942, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1718061-ed", "ovs_interfaceid": "b1718061-edf1-4077-b57f-0e54aea365a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 890.180828] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1f:27:16', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a316376e-2ef0-4b1e-b40c-10321ebd7e1a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b1718061-edf1-4077-b57f-0e54aea365a6', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 890.188365] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Creating folder: Project (1952dedf340f4b7ba0d57af6bba9a749). Parent ref: group-v294931. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 890.188651] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b4e03cd9-fa1e-4ac2-842b-eb63206eb789 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.198576] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Created folder: Project (1952dedf340f4b7ba0d57af6bba9a749) in parent group-v294931. [ 890.198751] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Creating folder: Instances. Parent ref: group-v294995. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 890.199262] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4ca60a79-869f-46a4-b5bb-d847bdeddd23 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.207509] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Created folder: Instances in parent group-v294995. [ 890.207750] env[62965]: DEBUG oslo.service.loopingcall [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 890.207950] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 890.208168] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-947f6924-9a05-4dcc-9094-c105a3868309 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.226243] env[62965]: DEBUG nova.compute.manager [req-c8deed39-e116-49ae-8926-6f7d3792dbce req-3075e3dc-c7c4-4388-b45d-11c2c9b420fc service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Received event network-vif-plugged-b1718061-edf1-4077-b57f-0e54aea365a6 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 890.226437] env[62965]: DEBUG oslo_concurrency.lockutils [req-c8deed39-e116-49ae-8926-6f7d3792dbce req-3075e3dc-c7c4-4388-b45d-11c2c9b420fc service nova] Acquiring lock "d038608c-9264-4fd5-b9ba-0653c98ec0e4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.226771] env[62965]: DEBUG oslo_concurrency.lockutils [req-c8deed39-e116-49ae-8926-6f7d3792dbce req-3075e3dc-c7c4-4388-b45d-11c2c9b420fc service nova] Lock "d038608c-9264-4fd5-b9ba-0653c98ec0e4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.226861] env[62965]: DEBUG oslo_concurrency.lockutils [req-c8deed39-e116-49ae-8926-6f7d3792dbce req-3075e3dc-c7c4-4388-b45d-11c2c9b420fc service nova] Lock "d038608c-9264-4fd5-b9ba-0653c98ec0e4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.226966] env[62965]: DEBUG nova.compute.manager [req-c8deed39-e116-49ae-8926-6f7d3792dbce req-3075e3dc-c7c4-4388-b45d-11c2c9b420fc service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] No waiting events found dispatching network-vif-plugged-b1718061-edf1-4077-b57f-0e54aea365a6 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 890.227135] env[62965]: WARNING nova.compute.manager [req-c8deed39-e116-49ae-8926-6f7d3792dbce req-3075e3dc-c7c4-4388-b45d-11c2c9b420fc service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Received unexpected event network-vif-plugged-b1718061-edf1-4077-b57f-0e54aea365a6 for instance with vm_state building and task_state spawning. [ 890.227295] env[62965]: DEBUG nova.compute.manager [req-c8deed39-e116-49ae-8926-6f7d3792dbce req-3075e3dc-c7c4-4388-b45d-11c2c9b420fc service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Received event network-changed-b1718061-edf1-4077-b57f-0e54aea365a6 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 890.227451] env[62965]: DEBUG nova.compute.manager [req-c8deed39-e116-49ae-8926-6f7d3792dbce req-3075e3dc-c7c4-4388-b45d-11c2c9b420fc service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Refreshing instance network info cache due to event network-changed-b1718061-edf1-4077-b57f-0e54aea365a6. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 890.227709] env[62965]: DEBUG oslo_concurrency.lockutils [req-c8deed39-e116-49ae-8926-6f7d3792dbce req-3075e3dc-c7c4-4388-b45d-11c2c9b420fc service nova] Acquiring lock "refresh_cache-d038608c-9264-4fd5-b9ba-0653c98ec0e4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.227943] env[62965]: DEBUG oslo_concurrency.lockutils [req-c8deed39-e116-49ae-8926-6f7d3792dbce req-3075e3dc-c7c4-4388-b45d-11c2c9b420fc service nova] Acquired lock "refresh_cache-d038608c-9264-4fd5-b9ba-0653c98ec0e4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.227999] env[62965]: DEBUG nova.network.neutron [req-c8deed39-e116-49ae-8926-6f7d3792dbce req-3075e3dc-c7c4-4388-b45d-11c2c9b420fc service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Refreshing network info cache for port b1718061-edf1-4077-b57f-0e54aea365a6 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 890.233282] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 890.233282] env[62965]: value = "task-1389937" [ 890.233282] env[62965]: _type = "Task" [ 890.233282] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.241444] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389937, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.304718] env[62965]: DEBUG oslo_vmware.api [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1389934, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.812844} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.305017] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 4fc543a1-34ce-40bb-9e80-72ab9098eb38/4fc543a1-34ce-40bb-9e80-72ab9098eb38.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 890.305281] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 890.305651] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-07fd4380-2133-4833-be6b-423d55368081 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.312421] env[62965]: DEBUG oslo_vmware.api [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 890.312421] env[62965]: value = "task-1389938" [ 890.312421] env[62965]: _type = "Task" [ 890.312421] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.320109] env[62965]: DEBUG oslo_vmware.api [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1389938, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.337788] env[62965]: DEBUG oslo_vmware.api [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5294853c-60e3-8804-6670-ba6daa616caa, 'name': SearchDatastore_Task, 'duration_secs': 0.060316} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.338105] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.338374] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 8c478341-9b50-45de-9ba3-321e802ed1f0/8c478341-9b50-45de-9ba3-321e802ed1f0.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 890.338637] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-90667857-a15c-424e-8019-170bb39d06b6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.345705] env[62965]: DEBUG oslo_vmware.api [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 890.345705] env[62965]: value = "task-1389939" [ 890.345705] env[62965]: _type = "Task" [ 890.345705] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.353134] env[62965]: DEBUG oslo_vmware.api [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1389939, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.545413] env[62965]: DEBUG nova.compute.manager [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 890.655572] env[62965]: DEBUG nova.scheduler.client.report [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 890.686408] env[62965]: DEBUG oslo_concurrency.lockutils [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.742455] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389937, 'name': CreateVM_Task} progress is 99%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.822981] env[62965]: DEBUG oslo_vmware.api [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1389938, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068918} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.823592] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 890.824541] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3ed8301-c871-4a72-a8be-8541ed2aca4c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.848569] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] 4fc543a1-34ce-40bb-9e80-72ab9098eb38/4fc543a1-34ce-40bb-9e80-72ab9098eb38.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 890.851177] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fe188200-c37a-46ae-8dbf-6f0464f92db3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.875661] env[62965]: DEBUG oslo_vmware.api [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1389939, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.477955} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.877281] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 8c478341-9b50-45de-9ba3-321e802ed1f0/8c478341-9b50-45de-9ba3-321e802ed1f0.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 890.877550] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 890.877865] env[62965]: DEBUG oslo_vmware.api [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 890.877865] env[62965]: value = "task-1389940" [ 890.877865] env[62965]: _type = "Task" [ 890.877865] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.878096] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-035aa997-8132-4680-9e2d-cf635f99310f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.888519] env[62965]: DEBUG oslo_vmware.api [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1389940, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.889876] env[62965]: DEBUG oslo_vmware.api [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 890.889876] env[62965]: value = "task-1389941" [ 890.889876] env[62965]: _type = "Task" [ 890.889876] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.897665] env[62965]: DEBUG oslo_vmware.api [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1389941, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.998124] env[62965]: DEBUG nova.network.neutron [req-c8deed39-e116-49ae-8926-6f7d3792dbce req-3075e3dc-c7c4-4388-b45d-11c2c9b420fc service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Updated VIF entry in instance network info cache for port b1718061-edf1-4077-b57f-0e54aea365a6. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 890.998686] env[62965]: DEBUG nova.network.neutron [req-c8deed39-e116-49ae-8926-6f7d3792dbce req-3075e3dc-c7c4-4388-b45d-11c2c9b420fc service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Updating instance_info_cache with network_info: [{"id": "b1718061-edf1-4077-b57f-0e54aea365a6", "address": "fa:16:3e:1f:27:16", "network": {"id": "a9699d8b-5e76-4666-a236-33ab8f22345e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1329540850-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1952dedf340f4b7ba0d57af6bba9a749", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a316376e-2ef0-4b1e-b40c-10321ebd7e1a", "external-id": "nsx-vlan-transportzone-942", "segmentation_id": 942, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1718061-ed", "ovs_interfaceid": "b1718061-edf1-4077-b57f-0e54aea365a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.064566] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.160969] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.246s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.161590] env[62965]: DEBUG nova.compute.manager [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 891.164334] env[62965]: DEBUG oslo_concurrency.lockutils [None req-331e5ade-da3c-4f0a-9a4b-a82bb76a893e tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.482s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.164524] env[62965]: DEBUG nova.objects.instance [None req-331e5ade-da3c-4f0a-9a4b-a82bb76a893e tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Lazy-loading 'resources' on Instance uuid b0d8f38a-86d3-466a-9b83-35145071dc21 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 891.243190] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389937, 'name': CreateVM_Task, 'duration_secs': 0.527251} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.243364] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 891.244085] env[62965]: DEBUG oslo_concurrency.lockutils [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.244253] env[62965]: DEBUG oslo_concurrency.lockutils [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.244794] env[62965]: DEBUG oslo_concurrency.lockutils [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 891.244867] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c278483a-a7a7-4bfc-99cd-83cd69f50d25 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.249366] env[62965]: DEBUG oslo_vmware.api [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Waiting for the task: (returnval){ [ 891.249366] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]525dc98b-e6dc-bddc-1557-43337d47ee01" [ 891.249366] env[62965]: _type = "Task" [ 891.249366] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.257622] env[62965]: DEBUG oslo_vmware.api [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]525dc98b-e6dc-bddc-1557-43337d47ee01, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.390265] env[62965]: DEBUG oslo_vmware.api [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1389940, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.397988] env[62965]: DEBUG oslo_vmware.api [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1389941, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.345585} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.398295] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 891.399304] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-196121d7-30af-4b30-9229-5935f4d4e465 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.421921] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] 8c478341-9b50-45de-9ba3-321e802ed1f0/8c478341-9b50-45de-9ba3-321e802ed1f0.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 891.423025] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-029a9562-9cc6-4e63-9dd9-65511d680950 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.441969] env[62965]: DEBUG oslo_vmware.api [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 891.441969] env[62965]: value = "task-1389942" [ 891.441969] env[62965]: _type = "Task" [ 891.441969] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.451016] env[62965]: DEBUG oslo_vmware.api [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1389942, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.504476] env[62965]: DEBUG oslo_concurrency.lockutils [req-c8deed39-e116-49ae-8926-6f7d3792dbce req-3075e3dc-c7c4-4388-b45d-11c2c9b420fc service nova] Releasing lock "refresh_cache-d038608c-9264-4fd5-b9ba-0653c98ec0e4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.668107] env[62965]: DEBUG nova.compute.utils [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 891.674775] env[62965]: DEBUG nova.compute.manager [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 891.674993] env[62965]: DEBUG nova.network.neutron [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 891.722456] env[62965]: DEBUG nova.policy [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8d4f4af6d0244ef9ae3022c5fafda65d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c9378a978bd945ce90f13c3e615bee64', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 891.766610] env[62965]: DEBUG oslo_vmware.api [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]525dc98b-e6dc-bddc-1557-43337d47ee01, 'name': SearchDatastore_Task, 'duration_secs': 0.053167} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.767054] env[62965]: DEBUG oslo_concurrency.lockutils [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.767402] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 891.767761] env[62965]: DEBUG oslo_concurrency.lockutils [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.767993] env[62965]: DEBUG oslo_concurrency.lockutils [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.768284] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 891.768674] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ae495132-ef71-43f4-9208-8686279c8f67 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.785613] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 891.785613] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 891.787033] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d919db03-a822-4a47-9e28-1e1d38d2903e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.794189] env[62965]: DEBUG oslo_vmware.api [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Waiting for the task: (returnval){ [ 891.794189] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5226bd41-0823-f84b-0bf1-7ac7ce3d1a8a" [ 891.794189] env[62965]: _type = "Task" [ 891.794189] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.813089] env[62965]: DEBUG oslo_vmware.api [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5226bd41-0823-f84b-0bf1-7ac7ce3d1a8a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.898316] env[62965]: DEBUG oslo_vmware.api [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1389940, 'name': ReconfigVM_Task, 'duration_secs': 0.922516} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.898504] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Reconfigured VM instance instance-0000004b to attach disk [datastore2] 4fc543a1-34ce-40bb-9e80-72ab9098eb38/4fc543a1-34ce-40bb-9e80-72ab9098eb38.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 891.899163] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-125bb1d2-06a3-4316-b232-6dd83a7a9d9a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.907068] env[62965]: DEBUG oslo_vmware.api [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 891.907068] env[62965]: value = "task-1389943" [ 891.907068] env[62965]: _type = "Task" [ 891.907068] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.915727] env[62965]: DEBUG oslo_vmware.api [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1389943, 'name': Rename_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.952404] env[62965]: DEBUG oslo_vmware.api [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1389942, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.970167] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee6991b9-daee-49b0-a5e5-c7256e7abf9a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.977852] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcf5c848-67f9-4238-851d-daec2959e001 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.009754] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74dbf03d-0ae0-4a17-baa0-50d583786ad7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.017652] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-288d6cd3-7446-4730-8cee-9387c50e1505 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.032344] env[62965]: DEBUG nova.compute.provider_tree [None req-331e5ade-da3c-4f0a-9a4b-a82bb76a893e tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 892.101541] env[62965]: DEBUG nova.network.neutron [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Successfully created port: 34d9279a-1a85-4e01-8a4f-b07cb77ec0e6 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 892.175525] env[62965]: DEBUG nova.compute.manager [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 892.307117] env[62965]: DEBUG oslo_vmware.api [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5226bd41-0823-f84b-0bf1-7ac7ce3d1a8a, 'name': SearchDatastore_Task, 'duration_secs': 0.025184} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.308174] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d51d903-062c-46cb-9dce-bef197319610 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.313496] env[62965]: DEBUG oslo_vmware.api [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Waiting for the task: (returnval){ [ 892.313496] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52648550-45e7-882f-bb55-5b90b0a456a7" [ 892.313496] env[62965]: _type = "Task" [ 892.313496] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.321396] env[62965]: DEBUG oslo_vmware.api [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52648550-45e7-882f-bb55-5b90b0a456a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.416363] env[62965]: DEBUG oslo_vmware.api [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1389943, 'name': Rename_Task} progress is 99%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.453136] env[62965]: DEBUG oslo_vmware.api [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1389942, 'name': ReconfigVM_Task, 'duration_secs': 0.620145} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.453136] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Reconfigured VM instance instance-00000046 to attach disk [datastore2] 8c478341-9b50-45de-9ba3-321e802ed1f0/8c478341-9b50-45de-9ba3-321e802ed1f0.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 892.453513] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2089a0ea-c100-405f-b391-74f978ab77d2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.460303] env[62965]: DEBUG oslo_vmware.api [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 892.460303] env[62965]: value = "task-1389944" [ 892.460303] env[62965]: _type = "Task" [ 892.460303] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.468643] env[62965]: DEBUG oslo_vmware.api [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1389944, 'name': Rename_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.536452] env[62965]: DEBUG nova.scheduler.client.report [None req-331e5ade-da3c-4f0a-9a4b-a82bb76a893e tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 892.824749] env[62965]: DEBUG oslo_vmware.api [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52648550-45e7-882f-bb55-5b90b0a456a7, 'name': SearchDatastore_Task, 'duration_secs': 0.055096} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.826171] env[62965]: DEBUG oslo_concurrency.lockutils [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.826171] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] d038608c-9264-4fd5-b9ba-0653c98ec0e4/d038608c-9264-4fd5-b9ba-0653c98ec0e4.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 892.826171] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-40208b5d-3609-44e5-b81e-2039524ea55b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.832591] env[62965]: DEBUG oslo_vmware.api [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Waiting for the task: (returnval){ [ 892.832591] env[62965]: value = "task-1389945" [ 892.832591] env[62965]: _type = "Task" [ 892.832591] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.841114] env[62965]: DEBUG oslo_vmware.api [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1389945, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.917360] env[62965]: DEBUG oslo_vmware.api [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1389943, 'name': Rename_Task} progress is 99%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.970210] env[62965]: DEBUG oslo_vmware.api [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1389944, 'name': Rename_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.041237] env[62965]: DEBUG oslo_concurrency.lockutils [None req-331e5ade-da3c-4f0a-9a4b-a82bb76a893e tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.877s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.044384] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.579s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.045066] env[62965]: INFO nova.compute.claims [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 893.066529] env[62965]: INFO nova.scheduler.client.report [None req-331e5ade-da3c-4f0a-9a4b-a82bb76a893e tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Deleted allocations for instance b0d8f38a-86d3-466a-9b83-35145071dc21 [ 893.185661] env[62965]: DEBUG nova.compute.manager [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 893.214129] env[62965]: DEBUG nova.virt.hardware [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 893.214425] env[62965]: DEBUG nova.virt.hardware [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 893.214584] env[62965]: DEBUG nova.virt.hardware [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 893.214773] env[62965]: DEBUG nova.virt.hardware [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 893.214922] env[62965]: DEBUG nova.virt.hardware [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 893.215127] env[62965]: DEBUG nova.virt.hardware [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 893.215305] env[62965]: DEBUG nova.virt.hardware [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 893.215465] env[62965]: DEBUG nova.virt.hardware [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 893.215643] env[62965]: DEBUG nova.virt.hardware [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 893.215796] env[62965]: DEBUG nova.virt.hardware [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 893.216066] env[62965]: DEBUG nova.virt.hardware [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 893.216956] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f2994ae-b504-4d8d-8a99-5c4b11c69068 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.227044] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3bdcd69-2893-4af1-a935-fa3edf612a0c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.343800] env[62965]: DEBUG oslo_vmware.api [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1389945, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.418476] env[62965]: DEBUG oslo_vmware.api [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1389943, 'name': Rename_Task, 'duration_secs': 1.259397} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.418857] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 893.419043] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4260aaa2-d0c5-4be0-8f09-1f39b915458e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.428651] env[62965]: DEBUG oslo_vmware.api [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 893.428651] env[62965]: value = "task-1389946" [ 893.428651] env[62965]: _type = "Task" [ 893.428651] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.440055] env[62965]: DEBUG oslo_vmware.api [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1389946, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.470096] env[62965]: DEBUG oslo_vmware.api [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1389944, 'name': Rename_Task, 'duration_secs': 0.885075} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.470313] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 893.470672] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6f255b24-40fa-4013-b92e-0955e72ebf74 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.477151] env[62965]: DEBUG oslo_vmware.api [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 893.477151] env[62965]: value = "task-1389947" [ 893.477151] env[62965]: _type = "Task" [ 893.477151] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.490243] env[62965]: DEBUG oslo_vmware.api [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1389947, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.575259] env[62965]: DEBUG oslo_concurrency.lockutils [None req-331e5ade-da3c-4f0a-9a4b-a82bb76a893e tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Lock "b0d8f38a-86d3-466a-9b83-35145071dc21" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.081s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.844353] env[62965]: DEBUG oslo_vmware.api [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1389945, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.643133} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.844656] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] d038608c-9264-4fd5-b9ba-0653c98ec0e4/d038608c-9264-4fd5-b9ba-0653c98ec0e4.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 893.844904] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 893.845201] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b3d4b2fb-ecd9-4ec8-bb2d-3e8c3548e9b7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.851256] env[62965]: DEBUG oslo_vmware.api [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Waiting for the task: (returnval){ [ 893.851256] env[62965]: value = "task-1389948" [ 893.851256] env[62965]: _type = "Task" [ 893.851256] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.859197] env[62965]: DEBUG oslo_vmware.api [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1389948, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.940791] env[62965]: DEBUG oslo_vmware.api [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1389946, 'name': PowerOnVM_Task} progress is 90%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.988490] env[62965]: DEBUG oslo_vmware.api [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1389947, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.275964] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93891f9c-1c69-457b-bf10-826b807e11b2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.283928] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3b23b66-e5c3-4d89-a9b4-6013e1e02bf2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.291195] env[62965]: DEBUG nova.compute.manager [req-fc315102-e5cc-4234-b11d-882fe9a27469 req-f5e94150-9a22-47ff-92ea-076e159065a6 service nova] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Received event network-vif-plugged-34d9279a-1a85-4e01-8a4f-b07cb77ec0e6 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 894.291494] env[62965]: DEBUG oslo_concurrency.lockutils [req-fc315102-e5cc-4234-b11d-882fe9a27469 req-f5e94150-9a22-47ff-92ea-076e159065a6 service nova] Acquiring lock "fa511b61-887e-46a1-aaf6-586d716c0fb5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.291785] env[62965]: DEBUG oslo_concurrency.lockutils [req-fc315102-e5cc-4234-b11d-882fe9a27469 req-f5e94150-9a22-47ff-92ea-076e159065a6 service nova] Lock "fa511b61-887e-46a1-aaf6-586d716c0fb5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.291991] env[62965]: DEBUG oslo_concurrency.lockutils [req-fc315102-e5cc-4234-b11d-882fe9a27469 req-f5e94150-9a22-47ff-92ea-076e159065a6 service nova] Lock "fa511b61-887e-46a1-aaf6-586d716c0fb5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.292241] env[62965]: DEBUG nova.compute.manager [req-fc315102-e5cc-4234-b11d-882fe9a27469 req-f5e94150-9a22-47ff-92ea-076e159065a6 service nova] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] No waiting events found dispatching network-vif-plugged-34d9279a-1a85-4e01-8a4f-b07cb77ec0e6 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 894.292470] env[62965]: WARNING nova.compute.manager [req-fc315102-e5cc-4234-b11d-882fe9a27469 req-f5e94150-9a22-47ff-92ea-076e159065a6 service nova] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Received unexpected event network-vif-plugged-34d9279a-1a85-4e01-8a4f-b07cb77ec0e6 for instance with vm_state building and task_state spawning. [ 894.322752] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cad044f7-5835-4b4e-bd4c-55b7a300054c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.330233] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-506883ce-414a-4e2e-9069-f136c16e8766 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.344218] env[62965]: DEBUG nova.compute.provider_tree [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 894.359043] env[62965]: DEBUG oslo_vmware.api [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1389948, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.267289} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.359426] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 894.360239] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f3176e5-5ce3-4d57-b01c-e16137f8ba2e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.394475] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] d038608c-9264-4fd5-b9ba-0653c98ec0e4/d038608c-9264-4fd5-b9ba-0653c98ec0e4.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 894.396323] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cda3df9c-8b2a-48eb-9283-1e56da192b63 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.419515] env[62965]: DEBUG oslo_vmware.api [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Waiting for the task: (returnval){ [ 894.419515] env[62965]: value = "task-1389949" [ 894.419515] env[62965]: _type = "Task" [ 894.419515] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.428250] env[62965]: DEBUG oslo_vmware.api [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1389949, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.438428] env[62965]: DEBUG oslo_vmware.api [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1389946, 'name': PowerOnVM_Task, 'duration_secs': 0.587698} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.438706] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 894.438963] env[62965]: INFO nova.compute.manager [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Took 13.11 seconds to spawn the instance on the hypervisor. [ 894.439164] env[62965]: DEBUG nova.compute.manager [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 894.439977] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4ef0e26-2022-4ef2-a6ca-dec52ef99c98 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.455551] env[62965]: DEBUG nova.network.neutron [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Successfully updated port: 34d9279a-1a85-4e01-8a4f-b07cb77ec0e6 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 894.488286] env[62965]: DEBUG oslo_vmware.api [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1389947, 'name': PowerOnVM_Task, 'duration_secs': 0.535443} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.488610] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 894.488835] env[62965]: DEBUG nova.compute.manager [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 894.490448] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-880a6c55-3ebc-4678-9578-3346e67ec8fd {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.847903] env[62965]: DEBUG nova.scheduler.client.report [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 894.930508] env[62965]: DEBUG oslo_vmware.api [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1389949, 'name': ReconfigVM_Task, 'duration_secs': 0.28546} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.930820] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Reconfigured VM instance instance-0000004c to attach disk [datastore2] d038608c-9264-4fd5-b9ba-0653c98ec0e4/d038608c-9264-4fd5-b9ba-0653c98ec0e4.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 894.931507] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-db8fdc13-e365-4de5-9092-11cad0102b3a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.938446] env[62965]: DEBUG oslo_vmware.api [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Waiting for the task: (returnval){ [ 894.938446] env[62965]: value = "task-1389950" [ 894.938446] env[62965]: _type = "Task" [ 894.938446] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.954929] env[62965]: DEBUG oslo_vmware.api [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1389950, 'name': Rename_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.957556] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Acquiring lock "refresh_cache-fa511b61-887e-46a1-aaf6-586d716c0fb5" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.957824] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Acquired lock "refresh_cache-fa511b61-887e-46a1-aaf6-586d716c0fb5" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.957998] env[62965]: DEBUG nova.network.neutron [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 894.960033] env[62965]: INFO nova.compute.manager [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Took 26.62 seconds to build instance. [ 895.006616] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.178714] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Acquiring lock "e0c4e188-a041-4aa7-95de-6e8718b17ade" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.178858] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Lock "e0c4e188-a041-4aa7-95de-6e8718b17ade" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.353362] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.310s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.354023] env[62965]: DEBUG nova.compute.manager [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 895.356518] env[62965]: DEBUG oslo_concurrency.lockutils [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.670s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.358088] env[62965]: INFO nova.compute.claims [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 895.448346] env[62965]: DEBUG oslo_vmware.api [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1389950, 'name': Rename_Task, 'duration_secs': 0.129396} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.448705] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 895.448973] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a628a93c-4ab7-4762-b2de-b68e7648410e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.454777] env[62965]: DEBUG oslo_vmware.api [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Waiting for the task: (returnval){ [ 895.454777] env[62965]: value = "task-1389951" [ 895.454777] env[62965]: _type = "Task" [ 895.454777] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.463455] env[62965]: DEBUG oslo_vmware.api [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1389951, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.463807] env[62965]: DEBUG oslo_concurrency.lockutils [None req-444789a1-ff6f-4829-9345-f19593c7ebff tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "4fc543a1-34ce-40bb-9e80-72ab9098eb38" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.046s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.492349] env[62965]: DEBUG nova.network.neutron [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 895.681401] env[62965]: DEBUG nova.compute.manager [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 895.717296] env[62965]: DEBUG nova.network.neutron [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Updating instance_info_cache with network_info: [{"id": "34d9279a-1a85-4e01-8a4f-b07cb77ec0e6", "address": "fa:16:3e:63:e5:94", "network": {"id": "b73a337f-00a6-4135-8e2d-96e3e0580e52", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-392444545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9378a978bd945ce90f13c3e615bee64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20641d67-1612-4b9c-8924-7a77df9c8e6d", "external-id": "nsx-vlan-transportzone-884", "segmentation_id": 884, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34d9279a-1a", "ovs_interfaceid": "34d9279a-1a85-4e01-8a4f-b07cb77ec0e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.863319] env[62965]: DEBUG nova.compute.utils [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 895.867895] env[62965]: DEBUG nova.compute.manager [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 895.868173] env[62965]: DEBUG nova.network.neutron [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 895.943982] env[62965]: DEBUG nova.policy [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ce0e7f3429454fd18d54cbb964dfb561', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e1c6e8403409438f9ff10235cd2533bc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 895.965836] env[62965]: DEBUG oslo_vmware.api [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1389951, 'name': PowerOnVM_Task, 'duration_secs': 0.443578} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.966166] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 895.966417] env[62965]: INFO nova.compute.manager [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Took 9.22 seconds to spawn the instance on the hypervisor. [ 895.966860] env[62965]: DEBUG nova.compute.manager [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 895.968513] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-650e2c67-416d-4392-a497-f1042c14de06 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.031698] env[62965]: DEBUG oslo_concurrency.lockutils [None req-322bb93e-e867-4da5-a1a8-6a5eae03a2a1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "8c478341-9b50-45de-9ba3-321e802ed1f0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.032577] env[62965]: DEBUG oslo_concurrency.lockutils [None req-322bb93e-e867-4da5-a1a8-6a5eae03a2a1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "8c478341-9b50-45de-9ba3-321e802ed1f0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.032961] env[62965]: DEBUG oslo_concurrency.lockutils [None req-322bb93e-e867-4da5-a1a8-6a5eae03a2a1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "8c478341-9b50-45de-9ba3-321e802ed1f0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.033194] env[62965]: DEBUG oslo_concurrency.lockutils [None req-322bb93e-e867-4da5-a1a8-6a5eae03a2a1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "8c478341-9b50-45de-9ba3-321e802ed1f0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.033386] env[62965]: DEBUG oslo_concurrency.lockutils [None req-322bb93e-e867-4da5-a1a8-6a5eae03a2a1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "8c478341-9b50-45de-9ba3-321e802ed1f0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.037245] env[62965]: INFO nova.compute.manager [None req-322bb93e-e867-4da5-a1a8-6a5eae03a2a1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Terminating instance [ 896.204684] env[62965]: DEBUG nova.network.neutron [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Successfully created port: ff5bbb06-f9d8-4d9c-8884-d85b3dbbb90e {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 896.207420] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.219751] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Releasing lock "refresh_cache-fa511b61-887e-46a1-aaf6-586d716c0fb5" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.220098] env[62965]: DEBUG nova.compute.manager [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Instance network_info: |[{"id": "34d9279a-1a85-4e01-8a4f-b07cb77ec0e6", "address": "fa:16:3e:63:e5:94", "network": {"id": "b73a337f-00a6-4135-8e2d-96e3e0580e52", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-392444545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9378a978bd945ce90f13c3e615bee64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20641d67-1612-4b9c-8924-7a77df9c8e6d", "external-id": "nsx-vlan-transportzone-884", "segmentation_id": 884, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34d9279a-1a", "ovs_interfaceid": "34d9279a-1a85-4e01-8a4f-b07cb77ec0e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 896.220522] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:63:e5:94', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '20641d67-1612-4b9c-8924-7a77df9c8e6d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '34d9279a-1a85-4e01-8a4f-b07cb77ec0e6', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 896.229351] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Creating folder: Project (c9378a978bd945ce90f13c3e615bee64). Parent ref: group-v294931. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 896.229351] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7623aebf-4566-405d-9140-a1402cb55400 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.244556] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Created folder: Project (c9378a978bd945ce90f13c3e615bee64) in parent group-v294931. [ 896.244753] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Creating folder: Instances. Parent ref: group-v294998. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 896.244999] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-35132cf3-9990-46d9-a1ed-f1b15bae84dd {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.253986] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Created folder: Instances in parent group-v294998. [ 896.254239] env[62965]: DEBUG oslo.service.loopingcall [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 896.254435] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 896.254639] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-71fb4f1e-2df5-4774-9977-b1c35f6a4f3f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.278672] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 896.278672] env[62965]: value = "task-1389954" [ 896.278672] env[62965]: _type = "Task" [ 896.278672] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.286359] env[62965]: DEBUG oslo_concurrency.lockutils [None req-57dfb0c1-c576-4f51-b7a4-f919972a1f07 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "4fc543a1-34ce-40bb-9e80-72ab9098eb38" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.286604] env[62965]: DEBUG oslo_concurrency.lockutils [None req-57dfb0c1-c576-4f51-b7a4-f919972a1f07 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "4fc543a1-34ce-40bb-9e80-72ab9098eb38" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.286808] env[62965]: DEBUG oslo_concurrency.lockutils [None req-57dfb0c1-c576-4f51-b7a4-f919972a1f07 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "4fc543a1-34ce-40bb-9e80-72ab9098eb38-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.286991] env[62965]: DEBUG oslo_concurrency.lockutils [None req-57dfb0c1-c576-4f51-b7a4-f919972a1f07 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "4fc543a1-34ce-40bb-9e80-72ab9098eb38-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.287365] env[62965]: DEBUG oslo_concurrency.lockutils [None req-57dfb0c1-c576-4f51-b7a4-f919972a1f07 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "4fc543a1-34ce-40bb-9e80-72ab9098eb38-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.289027] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389954, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.289467] env[62965]: INFO nova.compute.manager [None req-57dfb0c1-c576-4f51-b7a4-f919972a1f07 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Terminating instance [ 896.368693] env[62965]: DEBUG nova.compute.manager [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 896.452390] env[62965]: DEBUG nova.compute.manager [req-5f16df43-e16a-4f65-b0b6-0df605b0680c req-ecc1c331-9cca-485e-9bd1-0b6596cee210 service nova] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Received event network-changed-34d9279a-1a85-4e01-8a4f-b07cb77ec0e6 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 896.452591] env[62965]: DEBUG nova.compute.manager [req-5f16df43-e16a-4f65-b0b6-0df605b0680c req-ecc1c331-9cca-485e-9bd1-0b6596cee210 service nova] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Refreshing instance network info cache due to event network-changed-34d9279a-1a85-4e01-8a4f-b07cb77ec0e6. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 896.452804] env[62965]: DEBUG oslo_concurrency.lockutils [req-5f16df43-e16a-4f65-b0b6-0df605b0680c req-ecc1c331-9cca-485e-9bd1-0b6596cee210 service nova] Acquiring lock "refresh_cache-fa511b61-887e-46a1-aaf6-586d716c0fb5" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.452943] env[62965]: DEBUG oslo_concurrency.lockutils [req-5f16df43-e16a-4f65-b0b6-0df605b0680c req-ecc1c331-9cca-485e-9bd1-0b6596cee210 service nova] Acquired lock "refresh_cache-fa511b61-887e-46a1-aaf6-586d716c0fb5" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.453119] env[62965]: DEBUG nova.network.neutron [req-5f16df43-e16a-4f65-b0b6-0df605b0680c req-ecc1c331-9cca-485e-9bd1-0b6596cee210 service nova] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Refreshing network info cache for port 34d9279a-1a85-4e01-8a4f-b07cb77ec0e6 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 896.493806] env[62965]: INFO nova.compute.manager [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Took 26.67 seconds to build instance. [ 896.541632] env[62965]: DEBUG nova.compute.manager [None req-322bb93e-e867-4da5-a1a8-6a5eae03a2a1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 896.541913] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-322bb93e-e867-4da5-a1a8-6a5eae03a2a1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 896.543473] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6f6da8c-4bd2-462b-bee4-755099b95786 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.553949] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-322bb93e-e867-4da5-a1a8-6a5eae03a2a1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 896.557907] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b9d0f434-bfc0-4c49-9489-b103f65e4845 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.566744] env[62965]: DEBUG oslo_vmware.api [None req-322bb93e-e867-4da5-a1a8-6a5eae03a2a1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 896.566744] env[62965]: value = "task-1389955" [ 896.566744] env[62965]: _type = "Task" [ 896.566744] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.575522] env[62965]: DEBUG oslo_vmware.api [None req-322bb93e-e867-4da5-a1a8-6a5eae03a2a1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1389955, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.642477] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "fbbcbe60-92f3-497d-b190-520a2547b3a9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.642792] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "fbbcbe60-92f3-497d-b190-520a2547b3a9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.686557] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3d65a3e-fd94-4747-921e-cf0103109b62 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.700098] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34c96fa1-5c93-40b3-aa03-52756071e438 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.736196] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8fd5c3f-67c1-4684-89b7-fec4ac1152f0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.743211] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c0304a1-5863-43fb-a9fa-7201a59d0add {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.758654] env[62965]: DEBUG nova.compute.provider_tree [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 896.790810] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389954, 'name': CreateVM_Task, 'duration_secs': 0.334809} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.791076] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 896.791804] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.792014] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.792327] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 896.792968] env[62965]: DEBUG nova.compute.manager [None req-57dfb0c1-c576-4f51-b7a4-f919972a1f07 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 896.793190] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-57dfb0c1-c576-4f51-b7a4-f919972a1f07 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 896.793435] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-20469543-866f-48ee-9c4f-29be8c96a0e2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.795815] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cdba4b7-4c6f-45b4-9851-6731a5410657 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.802573] env[62965]: DEBUG oslo_vmware.api [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Waiting for the task: (returnval){ [ 896.802573] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52225160-f327-d371-7ba7-86bfb78c24c3" [ 896.802573] env[62965]: _type = "Task" [ 896.802573] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.805075] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-57dfb0c1-c576-4f51-b7a4-f919972a1f07 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 896.807957] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-096f4c5e-6d6b-48ac-ba93-a0a656098f2d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.818512] env[62965]: DEBUG oslo_vmware.api [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52225160-f327-d371-7ba7-86bfb78c24c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.819965] env[62965]: DEBUG oslo_vmware.api [None req-57dfb0c1-c576-4f51-b7a4-f919972a1f07 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 896.819965] env[62965]: value = "task-1389956" [ 896.819965] env[62965]: _type = "Task" [ 896.819965] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.827913] env[62965]: DEBUG oslo_vmware.api [None req-57dfb0c1-c576-4f51-b7a4-f919972a1f07 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1389956, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.000249] env[62965]: DEBUG oslo_concurrency.lockutils [None req-66ccfc1e-4625-4fed-bc8b-4161d900808d tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "d038608c-9264-4fd5-b9ba-0653c98ec0e4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.568s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.076291] env[62965]: DEBUG oslo_vmware.api [None req-322bb93e-e867-4da5-a1a8-6a5eae03a2a1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1389955, 'name': PowerOffVM_Task, 'duration_secs': 0.309418} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.076558] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-322bb93e-e867-4da5-a1a8-6a5eae03a2a1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 897.077316] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-322bb93e-e867-4da5-a1a8-6a5eae03a2a1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 897.077316] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a43dcf6d-3589-4930-afef-ff19a93fed07 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.129624] env[62965]: DEBUG nova.compute.manager [req-9e4c8808-46dd-4ef3-a7f9-b0a329b1f6a8 req-e7fbde1e-5a04-494d-84b7-22ec2b0efdcb service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Received event network-changed-b1718061-edf1-4077-b57f-0e54aea365a6 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 897.129624] env[62965]: DEBUG nova.compute.manager [req-9e4c8808-46dd-4ef3-a7f9-b0a329b1f6a8 req-e7fbde1e-5a04-494d-84b7-22ec2b0efdcb service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Refreshing instance network info cache due to event network-changed-b1718061-edf1-4077-b57f-0e54aea365a6. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 897.129624] env[62965]: DEBUG oslo_concurrency.lockutils [req-9e4c8808-46dd-4ef3-a7f9-b0a329b1f6a8 req-e7fbde1e-5a04-494d-84b7-22ec2b0efdcb service nova] Acquiring lock "refresh_cache-d038608c-9264-4fd5-b9ba-0653c98ec0e4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.129624] env[62965]: DEBUG oslo_concurrency.lockutils [req-9e4c8808-46dd-4ef3-a7f9-b0a329b1f6a8 req-e7fbde1e-5a04-494d-84b7-22ec2b0efdcb service nova] Acquired lock "refresh_cache-d038608c-9264-4fd5-b9ba-0653c98ec0e4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.129624] env[62965]: DEBUG nova.network.neutron [req-9e4c8808-46dd-4ef3-a7f9-b0a329b1f6a8 req-e7fbde1e-5a04-494d-84b7-22ec2b0efdcb service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Refreshing network info cache for port b1718061-edf1-4077-b57f-0e54aea365a6 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 897.146949] env[62965]: DEBUG nova.compute.manager [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 897.152921] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-322bb93e-e867-4da5-a1a8-6a5eae03a2a1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 897.153102] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-322bb93e-e867-4da5-a1a8-6a5eae03a2a1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Deleting contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 897.157019] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-322bb93e-e867-4da5-a1a8-6a5eae03a2a1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Deleting the datastore file [datastore2] 8c478341-9b50-45de-9ba3-321e802ed1f0 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 897.157019] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dddc6434-edf0-44c6-a70e-4a4b07d8f02e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.160887] env[62965]: DEBUG oslo_vmware.api [None req-322bb93e-e867-4da5-a1a8-6a5eae03a2a1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 897.160887] env[62965]: value = "task-1389958" [ 897.160887] env[62965]: _type = "Task" [ 897.160887] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.176579] env[62965]: DEBUG oslo_vmware.api [None req-322bb93e-e867-4da5-a1a8-6a5eae03a2a1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1389958, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.236074] env[62965]: DEBUG nova.network.neutron [req-5f16df43-e16a-4f65-b0b6-0df605b0680c req-ecc1c331-9cca-485e-9bd1-0b6596cee210 service nova] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Updated VIF entry in instance network info cache for port 34d9279a-1a85-4e01-8a4f-b07cb77ec0e6. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 897.236074] env[62965]: DEBUG nova.network.neutron [req-5f16df43-e16a-4f65-b0b6-0df605b0680c req-ecc1c331-9cca-485e-9bd1-0b6596cee210 service nova] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Updating instance_info_cache with network_info: [{"id": "34d9279a-1a85-4e01-8a4f-b07cb77ec0e6", "address": "fa:16:3e:63:e5:94", "network": {"id": "b73a337f-00a6-4135-8e2d-96e3e0580e52", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-392444545-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c9378a978bd945ce90f13c3e615bee64", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20641d67-1612-4b9c-8924-7a77df9c8e6d", "external-id": "nsx-vlan-transportzone-884", "segmentation_id": 884, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap34d9279a-1a", "ovs_interfaceid": "34d9279a-1a85-4e01-8a4f-b07cb77ec0e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.263736] env[62965]: DEBUG nova.scheduler.client.report [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 897.315069] env[62965]: DEBUG oslo_vmware.api [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52225160-f327-d371-7ba7-86bfb78c24c3, 'name': SearchDatastore_Task, 'duration_secs': 0.012156} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.315385] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.315694] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 897.315873] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.316047] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.316239] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 897.316493] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1171cc77-f2f2-471d-9108-d66ea4d895da {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.328147] env[62965]: DEBUG oslo_vmware.api [None req-57dfb0c1-c576-4f51-b7a4-f919972a1f07 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1389956, 'name': PowerOffVM_Task, 'duration_secs': 0.403606} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.328389] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-57dfb0c1-c576-4f51-b7a4-f919972a1f07 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 897.328574] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-57dfb0c1-c576-4f51-b7a4-f919972a1f07 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 897.328844] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e0b63e78-d99a-4534-85c4-e51ffaf5d362 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.333250] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 897.333426] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 897.334273] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-32951414-22f9-4c78-aeb3-3e47deb5187c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.341165] env[62965]: DEBUG oslo_vmware.api [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Waiting for the task: (returnval){ [ 897.341165] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52f8ef49-6d0a-b836-c47c-4b7ddd5e60b3" [ 897.341165] env[62965]: _type = "Task" [ 897.341165] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.351580] env[62965]: DEBUG oslo_vmware.api [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52f8ef49-6d0a-b836-c47c-4b7ddd5e60b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.381235] env[62965]: DEBUG nova.compute.manager [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 897.394676] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-57dfb0c1-c576-4f51-b7a4-f919972a1f07 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 897.395052] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-57dfb0c1-c576-4f51-b7a4-f919972a1f07 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Deleting contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 897.395386] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-57dfb0c1-c576-4f51-b7a4-f919972a1f07 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Deleting the datastore file [datastore2] 4fc543a1-34ce-40bb-9e80-72ab9098eb38 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 897.395706] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2cabe7d5-07f8-4165-a2ae-9e187fb5f4fc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.405610] env[62965]: DEBUG oslo_vmware.api [None req-57dfb0c1-c576-4f51-b7a4-f919972a1f07 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 897.405610] env[62965]: value = "task-1389960" [ 897.405610] env[62965]: _type = "Task" [ 897.405610] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.413066] env[62965]: DEBUG nova.virt.hardware [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='1d9516733bcaf8c871f57bfaa596f6e9',container_format='bare',created_at=2024-10-31T12:36:36Z,direct_url=,disk_format='vmdk',id=7e436a80-d1ff-4c42-847a-9538961b1a6a,min_disk=1,min_ram=0,name='tempest-test-snap-2101270418',owner='e1c6e8403409438f9ff10235cd2533bc',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-31T12:36:53Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 897.413066] env[62965]: DEBUG nova.virt.hardware [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 897.413066] env[62965]: DEBUG nova.virt.hardware [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 897.413298] env[62965]: DEBUG nova.virt.hardware [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 897.413351] env[62965]: DEBUG nova.virt.hardware [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 897.413591] env[62965]: DEBUG nova.virt.hardware [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 897.413809] env[62965]: DEBUG nova.virt.hardware [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 897.414592] env[62965]: DEBUG nova.virt.hardware [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 897.414592] env[62965]: DEBUG nova.virt.hardware [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 897.414592] env[62965]: DEBUG nova.virt.hardware [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 897.414770] env[62965]: DEBUG nova.virt.hardware [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 897.416039] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-231a3aa0-b9b7-4e6d-bd97-104baa53cd58 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.424066] env[62965]: DEBUG oslo_vmware.api [None req-57dfb0c1-c576-4f51-b7a4-f919972a1f07 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1389960, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.427289] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d7f17b3-f9fd-48c2-887f-61dddbded647 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.669059] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.673135] env[62965]: DEBUG oslo_vmware.api [None req-322bb93e-e867-4da5-a1a8-6a5eae03a2a1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1389958, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.283814} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.673421] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-322bb93e-e867-4da5-a1a8-6a5eae03a2a1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 897.673601] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-322bb93e-e867-4da5-a1a8-6a5eae03a2a1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Deleted contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 897.673868] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-322bb93e-e867-4da5-a1a8-6a5eae03a2a1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 897.674128] env[62965]: INFO nova.compute.manager [None req-322bb93e-e867-4da5-a1a8-6a5eae03a2a1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Took 1.13 seconds to destroy the instance on the hypervisor. [ 897.674390] env[62965]: DEBUG oslo.service.loopingcall [None req-322bb93e-e867-4da5-a1a8-6a5eae03a2a1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 897.674588] env[62965]: DEBUG nova.compute.manager [-] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 897.674681] env[62965]: DEBUG nova.network.neutron [-] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 897.740039] env[62965]: DEBUG oslo_concurrency.lockutils [req-5f16df43-e16a-4f65-b0b6-0df605b0680c req-ecc1c331-9cca-485e-9bd1-0b6596cee210 service nova] Releasing lock "refresh_cache-fa511b61-887e-46a1-aaf6-586d716c0fb5" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.766493] env[62965]: DEBUG oslo_concurrency.lockutils [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.410s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.767039] env[62965]: DEBUG nova.compute.manager [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 897.769888] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.705s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.771332] env[62965]: INFO nova.compute.claims [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 897.858561] env[62965]: DEBUG oslo_vmware.api [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52f8ef49-6d0a-b836-c47c-4b7ddd5e60b3, 'name': SearchDatastore_Task, 'duration_secs': 0.018152} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.859443] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1725ec7-800e-48e3-91ea-3cac40f3249f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.865034] env[62965]: DEBUG oslo_vmware.api [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Waiting for the task: (returnval){ [ 897.865034] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52927727-3699-b762-f5d0-664c5c5bfde8" [ 897.865034] env[62965]: _type = "Task" [ 897.865034] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.875161] env[62965]: DEBUG oslo_vmware.api [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52927727-3699-b762-f5d0-664c5c5bfde8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.916227] env[62965]: DEBUG oslo_vmware.api [None req-57dfb0c1-c576-4f51-b7a4-f919972a1f07 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1389960, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.449499} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.916227] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-57dfb0c1-c576-4f51-b7a4-f919972a1f07 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 897.916227] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-57dfb0c1-c576-4f51-b7a4-f919972a1f07 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Deleted contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 897.916227] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-57dfb0c1-c576-4f51-b7a4-f919972a1f07 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 897.916708] env[62965]: INFO nova.compute.manager [None req-57dfb0c1-c576-4f51-b7a4-f919972a1f07 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Took 1.12 seconds to destroy the instance on the hypervisor. [ 897.917075] env[62965]: DEBUG oslo.service.loopingcall [None req-57dfb0c1-c576-4f51-b7a4-f919972a1f07 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 897.917430] env[62965]: DEBUG nova.compute.manager [-] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 897.917614] env[62965]: DEBUG nova.network.neutron [-] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 898.039152] env[62965]: DEBUG nova.network.neutron [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Successfully updated port: ff5bbb06-f9d8-4d9c-8884-d85b3dbbb90e {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 898.057130] env[62965]: DEBUG nova.network.neutron [req-9e4c8808-46dd-4ef3-a7f9-b0a329b1f6a8 req-e7fbde1e-5a04-494d-84b7-22ec2b0efdcb service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Updated VIF entry in instance network info cache for port b1718061-edf1-4077-b57f-0e54aea365a6. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 898.057698] env[62965]: DEBUG nova.network.neutron [req-9e4c8808-46dd-4ef3-a7f9-b0a329b1f6a8 req-e7fbde1e-5a04-494d-84b7-22ec2b0efdcb service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Updating instance_info_cache with network_info: [{"id": "b1718061-edf1-4077-b57f-0e54aea365a6", "address": "fa:16:3e:1f:27:16", "network": {"id": "a9699d8b-5e76-4666-a236-33ab8f22345e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1329540850-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1952dedf340f4b7ba0d57af6bba9a749", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a316376e-2ef0-4b1e-b40c-10321ebd7e1a", "external-id": "nsx-vlan-transportzone-942", "segmentation_id": 942, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1718061-ed", "ovs_interfaceid": "b1718061-edf1-4077-b57f-0e54aea365a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.108674] env[62965]: DEBUG oslo_concurrency.lockutils [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "00f24976-2f0e-4a2f-8f2e-012b0fce479f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.108674] env[62965]: DEBUG oslo_concurrency.lockutils [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "00f24976-2f0e-4a2f-8f2e-012b0fce479f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.276742] env[62965]: DEBUG nova.compute.utils [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 898.281837] env[62965]: DEBUG nova.compute.manager [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 898.282362] env[62965]: DEBUG nova.network.neutron [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 898.325893] env[62965]: DEBUG nova.policy [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '97ee4fdb7cdb45829c7fe694e883da8c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1ed71548c085499981fa1b102d660368', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 898.380019] env[62965]: DEBUG oslo_vmware.api [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52927727-3699-b762-f5d0-664c5c5bfde8, 'name': SearchDatastore_Task, 'duration_secs': 0.02895} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.380019] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.380019] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] fa511b61-887e-46a1-aaf6-586d716c0fb5/fa511b61-887e-46a1-aaf6-586d716c0fb5.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 898.380019] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1d40827a-8d99-42f4-a552-778b0bef8e54 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.386765] env[62965]: DEBUG oslo_vmware.api [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Waiting for the task: (returnval){ [ 898.386765] env[62965]: value = "task-1389961" [ 898.386765] env[62965]: _type = "Task" [ 898.386765] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.396285] env[62965]: DEBUG oslo_vmware.api [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Task: {'id': task-1389961, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.486674] env[62965]: DEBUG nova.compute.manager [req-7e0fc88e-5a09-4986-9c87-e122381a947a req-ceee8de3-a96b-4dae-a77f-93e06abfa2d5 service nova] [instance: f82d86ec-3364-449d-be54-35627082c02b] Received event network-vif-plugged-ff5bbb06-f9d8-4d9c-8884-d85b3dbbb90e {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 898.486987] env[62965]: DEBUG oslo_concurrency.lockutils [req-7e0fc88e-5a09-4986-9c87-e122381a947a req-ceee8de3-a96b-4dae-a77f-93e06abfa2d5 service nova] Acquiring lock "f82d86ec-3364-449d-be54-35627082c02b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.487290] env[62965]: DEBUG oslo_concurrency.lockutils [req-7e0fc88e-5a09-4986-9c87-e122381a947a req-ceee8de3-a96b-4dae-a77f-93e06abfa2d5 service nova] Lock "f82d86ec-3364-449d-be54-35627082c02b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.487478] env[62965]: DEBUG oslo_concurrency.lockutils [req-7e0fc88e-5a09-4986-9c87-e122381a947a req-ceee8de3-a96b-4dae-a77f-93e06abfa2d5 service nova] Lock "f82d86ec-3364-449d-be54-35627082c02b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.487681] env[62965]: DEBUG nova.compute.manager [req-7e0fc88e-5a09-4986-9c87-e122381a947a req-ceee8de3-a96b-4dae-a77f-93e06abfa2d5 service nova] [instance: f82d86ec-3364-449d-be54-35627082c02b] No waiting events found dispatching network-vif-plugged-ff5bbb06-f9d8-4d9c-8884-d85b3dbbb90e {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 898.487826] env[62965]: WARNING nova.compute.manager [req-7e0fc88e-5a09-4986-9c87-e122381a947a req-ceee8de3-a96b-4dae-a77f-93e06abfa2d5 service nova] [instance: f82d86ec-3364-449d-be54-35627082c02b] Received unexpected event network-vif-plugged-ff5bbb06-f9d8-4d9c-8884-d85b3dbbb90e for instance with vm_state building and task_state spawning. [ 898.487994] env[62965]: DEBUG nova.compute.manager [req-7e0fc88e-5a09-4986-9c87-e122381a947a req-ceee8de3-a96b-4dae-a77f-93e06abfa2d5 service nova] [instance: f82d86ec-3364-449d-be54-35627082c02b] Received event network-changed-ff5bbb06-f9d8-4d9c-8884-d85b3dbbb90e {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 898.488588] env[62965]: DEBUG nova.compute.manager [req-7e0fc88e-5a09-4986-9c87-e122381a947a req-ceee8de3-a96b-4dae-a77f-93e06abfa2d5 service nova] [instance: f82d86ec-3364-449d-be54-35627082c02b] Refreshing instance network info cache due to event network-changed-ff5bbb06-f9d8-4d9c-8884-d85b3dbbb90e. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 898.488588] env[62965]: DEBUG oslo_concurrency.lockutils [req-7e0fc88e-5a09-4986-9c87-e122381a947a req-ceee8de3-a96b-4dae-a77f-93e06abfa2d5 service nova] Acquiring lock "refresh_cache-f82d86ec-3364-449d-be54-35627082c02b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.488588] env[62965]: DEBUG oslo_concurrency.lockutils [req-7e0fc88e-5a09-4986-9c87-e122381a947a req-ceee8de3-a96b-4dae-a77f-93e06abfa2d5 service nova] Acquired lock "refresh_cache-f82d86ec-3364-449d-be54-35627082c02b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.488743] env[62965]: DEBUG nova.network.neutron [req-7e0fc88e-5a09-4986-9c87-e122381a947a req-ceee8de3-a96b-4dae-a77f-93e06abfa2d5 service nova] [instance: f82d86ec-3364-449d-be54-35627082c02b] Refreshing network info cache for port ff5bbb06-f9d8-4d9c-8884-d85b3dbbb90e {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 898.546163] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquiring lock "refresh_cache-f82d86ec-3364-449d-be54-35627082c02b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.560952] env[62965]: DEBUG oslo_concurrency.lockutils [req-9e4c8808-46dd-4ef3-a7f9-b0a329b1f6a8 req-e7fbde1e-5a04-494d-84b7-22ec2b0efdcb service nova] Releasing lock "refresh_cache-d038608c-9264-4fd5-b9ba-0653c98ec0e4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.603931] env[62965]: DEBUG nova.network.neutron [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Successfully created port: 6205e686-0f1c-46bf-a1e7-88f0fd8ee5e5 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 898.613038] env[62965]: DEBUG nova.compute.manager [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 898.664730] env[62965]: DEBUG nova.network.neutron [-] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.778914] env[62965]: DEBUG nova.network.neutron [-] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.789595] env[62965]: DEBUG nova.compute.manager [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 898.904028] env[62965]: DEBUG oslo_vmware.api [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Task: {'id': task-1389961, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.033127] env[62965]: DEBUG nova.network.neutron [req-7e0fc88e-5a09-4986-9c87-e122381a947a req-ceee8de3-a96b-4dae-a77f-93e06abfa2d5 service nova] [instance: f82d86ec-3364-449d-be54-35627082c02b] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 899.081049] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3d842dc-c945-4e1c-9b8a-2e053cca466b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.088321] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a75370a-0f71-49be-ba43-b9920ed6364f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.122800] env[62965]: DEBUG nova.network.neutron [req-7e0fc88e-5a09-4986-9c87-e122381a947a req-ceee8de3-a96b-4dae-a77f-93e06abfa2d5 service nova] [instance: f82d86ec-3364-449d-be54-35627082c02b] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.124499] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bb0d10b-0631-4087-b893-68bf5da1652e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.134535] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccbd6c5d-a73f-43a9-9589-1ea3482d8627 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.140036] env[62965]: DEBUG oslo_concurrency.lockutils [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.150464] env[62965]: DEBUG nova.compute.provider_tree [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 899.156053] env[62965]: DEBUG nova.compute.manager [req-2055b376-5a30-4fea-ab43-4e32640687b3 req-b3726032-fa53-4a3d-838a-e75ab797fbde service nova] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Received event network-vif-deleted-c0f50257-3550-4368-8ad1-aa07da3d6e67 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 899.156269] env[62965]: DEBUG nova.compute.manager [req-2055b376-5a30-4fea-ab43-4e32640687b3 req-b3726032-fa53-4a3d-838a-e75ab797fbde service nova] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Received event network-vif-deleted-4ab65d3b-4168-447e-b02a-0205186fa185 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 899.167554] env[62965]: INFO nova.compute.manager [-] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Took 1.49 seconds to deallocate network for instance. [ 899.284186] env[62965]: INFO nova.compute.manager [-] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Took 1.37 seconds to deallocate network for instance. [ 899.401203] env[62965]: DEBUG oslo_vmware.api [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Task: {'id': task-1389961, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.626445} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.401624] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] fa511b61-887e-46a1-aaf6-586d716c0fb5/fa511b61-887e-46a1-aaf6-586d716c0fb5.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 899.401907] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 899.402260] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3ae95b75-1096-48c1-8c19-a47217095331 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.411288] env[62965]: DEBUG oslo_vmware.api [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Waiting for the task: (returnval){ [ 899.411288] env[62965]: value = "task-1389962" [ 899.411288] env[62965]: _type = "Task" [ 899.411288] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.418872] env[62965]: DEBUG oslo_vmware.api [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Task: {'id': task-1389962, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.628581] env[62965]: DEBUG oslo_concurrency.lockutils [req-7e0fc88e-5a09-4986-9c87-e122381a947a req-ceee8de3-a96b-4dae-a77f-93e06abfa2d5 service nova] Releasing lock "refresh_cache-f82d86ec-3364-449d-be54-35627082c02b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.629011] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquired lock "refresh_cache-f82d86ec-3364-449d-be54-35627082c02b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.629204] env[62965]: DEBUG nova.network.neutron [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 899.653327] env[62965]: DEBUG nova.scheduler.client.report [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 899.673743] env[62965]: DEBUG oslo_concurrency.lockutils [None req-322bb93e-e867-4da5-a1a8-6a5eae03a2a1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.790901] env[62965]: DEBUG oslo_concurrency.lockutils [None req-57dfb0c1-c576-4f51-b7a4-f919972a1f07 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.799499] env[62965]: DEBUG nova.compute.manager [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 899.823784] env[62965]: DEBUG nova.virt.hardware [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 899.824047] env[62965]: DEBUG nova.virt.hardware [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 899.824216] env[62965]: DEBUG nova.virt.hardware [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 899.824406] env[62965]: DEBUG nova.virt.hardware [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 899.824553] env[62965]: DEBUG nova.virt.hardware [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 899.824699] env[62965]: DEBUG nova.virt.hardware [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 899.824915] env[62965]: DEBUG nova.virt.hardware [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 899.825096] env[62965]: DEBUG nova.virt.hardware [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 899.825275] env[62965]: DEBUG nova.virt.hardware [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 899.825437] env[62965]: DEBUG nova.virt.hardware [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 899.825612] env[62965]: DEBUG nova.virt.hardware [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 899.826492] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66d5c1b6-b1bf-4218-8c35-fea52ee14475 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.834325] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-148ed5fd-a110-4c69-aea6-f12f53d43f75 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.920599] env[62965]: DEBUG oslo_vmware.api [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Task: {'id': task-1389962, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066355} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.920934] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 899.921737] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed05c249-e7f5-448f-87c0-6321bda20de1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.944555] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] fa511b61-887e-46a1-aaf6-586d716c0fb5/fa511b61-887e-46a1-aaf6-586d716c0fb5.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 899.944838] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ed989b6a-8eda-4088-9d4d-4403b06a1bf5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.964518] env[62965]: DEBUG oslo_vmware.api [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Waiting for the task: (returnval){ [ 899.964518] env[62965]: value = "task-1389963" [ 899.964518] env[62965]: _type = "Task" [ 899.964518] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.972356] env[62965]: DEBUG oslo_vmware.api [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Task: {'id': task-1389963, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.158175] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.388s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.158728] env[62965]: DEBUG nova.compute.manager [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 900.161336] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 5.155s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.161538] env[62965]: DEBUG nova.objects.instance [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62965) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 900.174519] env[62965]: DEBUG nova.network.neutron [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 900.181607] env[62965]: DEBUG nova.network.neutron [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Successfully updated port: 6205e686-0f1c-46bf-a1e7-88f0fd8ee5e5 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 900.337257] env[62965]: DEBUG nova.network.neutron [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Updating instance_info_cache with network_info: [{"id": "ff5bbb06-f9d8-4d9c-8884-d85b3dbbb90e", "address": "fa:16:3e:d8:77:25", "network": {"id": "51f79fa0-8980-40d2-9255-a8da5b8fc5d7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-909239894-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1c6e8403409438f9ff10235cd2533bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff5bbb06-f9", "ovs_interfaceid": "ff5bbb06-f9d8-4d9c-8884-d85b3dbbb90e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.474682] env[62965]: DEBUG oslo_vmware.api [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Task: {'id': task-1389963, 'name': ReconfigVM_Task, 'duration_secs': 0.2979} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.474885] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Reconfigured VM instance instance-0000004d to attach disk [datastore2] fa511b61-887e-46a1-aaf6-586d716c0fb5/fa511b61-887e-46a1-aaf6-586d716c0fb5.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 900.475681] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-27bec57f-e698-4d50-8af4-8809bbe7aeed {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.482175] env[62965]: DEBUG oslo_vmware.api [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Waiting for the task: (returnval){ [ 900.482175] env[62965]: value = "task-1389964" [ 900.482175] env[62965]: _type = "Task" [ 900.482175] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.489626] env[62965]: DEBUG oslo_vmware.api [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Task: {'id': task-1389964, 'name': Rename_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.666189] env[62965]: DEBUG nova.compute.utils [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 900.670567] env[62965]: DEBUG nova.compute.manager [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 900.670783] env[62965]: DEBUG nova.network.neutron [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 900.685356] env[62965]: DEBUG oslo_concurrency.lockutils [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "refresh_cache-892b1302-f536-4dcd-9227-a545ae1ac731" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.685356] env[62965]: DEBUG oslo_concurrency.lockutils [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquired lock "refresh_cache-892b1302-f536-4dcd-9227-a545ae1ac731" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.685356] env[62965]: DEBUG nova.network.neutron [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 900.709204] env[62965]: DEBUG nova.policy [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f3c1bb4b1cd34098acea01a7656c0a57', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '89fd5275b2064288b44e79365f6f0271', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 900.840326] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Releasing lock "refresh_cache-f82d86ec-3364-449d-be54-35627082c02b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.841039] env[62965]: DEBUG nova.compute.manager [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Instance network_info: |[{"id": "ff5bbb06-f9d8-4d9c-8884-d85b3dbbb90e", "address": "fa:16:3e:d8:77:25", "network": {"id": "51f79fa0-8980-40d2-9255-a8da5b8fc5d7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-909239894-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1c6e8403409438f9ff10235cd2533bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff5bbb06-f9", "ovs_interfaceid": "ff5bbb06-f9d8-4d9c-8884-d85b3dbbb90e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 900.841218] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d8:77:25', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c6934071-bf85-4591-9c7d-55c7ea131262', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ff5bbb06-f9d8-4d9c-8884-d85b3dbbb90e', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 900.848555] env[62965]: DEBUG oslo.service.loopingcall [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 900.848774] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f82d86ec-3364-449d-be54-35627082c02b] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 900.848999] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c162afcd-b6d5-4032-8c21-2e363ec71285 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.870183] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 900.870183] env[62965]: value = "task-1389965" [ 900.870183] env[62965]: _type = "Task" [ 900.870183] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.877871] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389965, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.949978] env[62965]: DEBUG nova.network.neutron [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Successfully created port: 838a71f4-1d2e-4ffd-8ba7-a09d951c4a66 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 900.991752] env[62965]: DEBUG oslo_vmware.api [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Task: {'id': task-1389964, 'name': Rename_Task, 'duration_secs': 0.147335} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.991752] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 900.992529] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-32138d80-543f-4498-83ee-b61e40e6d63c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.999170] env[62965]: DEBUG oslo_vmware.api [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Waiting for the task: (returnval){ [ 900.999170] env[62965]: value = "task-1389966" [ 900.999170] env[62965]: _type = "Task" [ 900.999170] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.007474] env[62965]: DEBUG oslo_vmware.api [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Task: {'id': task-1389966, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.173771] env[62965]: DEBUG nova.compute.manager [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 901.177159] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e69ff718-714b-4f3b-a1b9-64c673640fe5 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.178220] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.971s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.180023] env[62965]: INFO nova.compute.claims [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 901.243710] env[62965]: DEBUG nova.network.neutron [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 901.338353] env[62965]: DEBUG nova.compute.manager [req-d2d5db2c-94f0-48fa-a429-de5175cdf0e6 req-b50756f7-5b12-4311-b66d-47fd1a91c32c service nova] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Received event network-vif-plugged-6205e686-0f1c-46bf-a1e7-88f0fd8ee5e5 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 901.338591] env[62965]: DEBUG oslo_concurrency.lockutils [req-d2d5db2c-94f0-48fa-a429-de5175cdf0e6 req-b50756f7-5b12-4311-b66d-47fd1a91c32c service nova] Acquiring lock "892b1302-f536-4dcd-9227-a545ae1ac731-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.338802] env[62965]: DEBUG oslo_concurrency.lockutils [req-d2d5db2c-94f0-48fa-a429-de5175cdf0e6 req-b50756f7-5b12-4311-b66d-47fd1a91c32c service nova] Lock "892b1302-f536-4dcd-9227-a545ae1ac731-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.338978] env[62965]: DEBUG oslo_concurrency.lockutils [req-d2d5db2c-94f0-48fa-a429-de5175cdf0e6 req-b50756f7-5b12-4311-b66d-47fd1a91c32c service nova] Lock "892b1302-f536-4dcd-9227-a545ae1ac731-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.339491] env[62965]: DEBUG nova.compute.manager [req-d2d5db2c-94f0-48fa-a429-de5175cdf0e6 req-b50756f7-5b12-4311-b66d-47fd1a91c32c service nova] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] No waiting events found dispatching network-vif-plugged-6205e686-0f1c-46bf-a1e7-88f0fd8ee5e5 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 901.339680] env[62965]: WARNING nova.compute.manager [req-d2d5db2c-94f0-48fa-a429-de5175cdf0e6 req-b50756f7-5b12-4311-b66d-47fd1a91c32c service nova] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Received unexpected event network-vif-plugged-6205e686-0f1c-46bf-a1e7-88f0fd8ee5e5 for instance with vm_state building and task_state spawning. [ 901.339906] env[62965]: DEBUG nova.compute.manager [req-d2d5db2c-94f0-48fa-a429-de5175cdf0e6 req-b50756f7-5b12-4311-b66d-47fd1a91c32c service nova] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Received event network-changed-6205e686-0f1c-46bf-a1e7-88f0fd8ee5e5 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 901.340058] env[62965]: DEBUG nova.compute.manager [req-d2d5db2c-94f0-48fa-a429-de5175cdf0e6 req-b50756f7-5b12-4311-b66d-47fd1a91c32c service nova] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Refreshing instance network info cache due to event network-changed-6205e686-0f1c-46bf-a1e7-88f0fd8ee5e5. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 901.340235] env[62965]: DEBUG oslo_concurrency.lockutils [req-d2d5db2c-94f0-48fa-a429-de5175cdf0e6 req-b50756f7-5b12-4311-b66d-47fd1a91c32c service nova] Acquiring lock "refresh_cache-892b1302-f536-4dcd-9227-a545ae1ac731" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.383827] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389965, 'name': CreateVM_Task, 'duration_secs': 0.39587} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.384215] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f82d86ec-3364-449d-be54-35627082c02b] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 901.385150] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7e436a80-d1ff-4c42-847a-9538961b1a6a" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.385344] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7e436a80-d1ff-4c42-847a-9538961b1a6a" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.385802] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/7e436a80-d1ff-4c42-847a-9538961b1a6a" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 901.386768] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3845286e-95b6-45e1-aa06-8937342401a2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.392525] env[62965]: DEBUG oslo_vmware.api [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 901.392525] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52501a2e-ef05-b7d7-fd20-db83597fe6fb" [ 901.392525] env[62965]: _type = "Task" [ 901.392525] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.401772] env[62965]: DEBUG oslo_vmware.api [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52501a2e-ef05-b7d7-fd20-db83597fe6fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.403925] env[62965]: DEBUG nova.network.neutron [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Updating instance_info_cache with network_info: [{"id": "6205e686-0f1c-46bf-a1e7-88f0fd8ee5e5", "address": "fa:16:3e:99:64:c5", "network": {"id": "a0a5e605-d580-4b7b-b3c0-5c7395bd5a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1679618017-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ed71548c085499981fa1b102d660368", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6205e686-0f", "ovs_interfaceid": "6205e686-0f1c-46bf-a1e7-88f0fd8ee5e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.508047] env[62965]: DEBUG oslo_vmware.api [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Task: {'id': task-1389966, 'name': PowerOnVM_Task, 'duration_secs': 0.462664} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.508333] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 901.508539] env[62965]: INFO nova.compute.manager [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Took 8.32 seconds to spawn the instance on the hypervisor. [ 901.508719] env[62965]: DEBUG nova.compute.manager [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 901.509517] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8957ab2-b8d1-483f-899b-40cf03b1cf70 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.902982] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7e436a80-d1ff-4c42-847a-9538961b1a6a" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.903273] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Processing image 7e436a80-d1ff-4c42-847a-9538961b1a6a {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 901.903509] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/7e436a80-d1ff-4c42-847a-9538961b1a6a/7e436a80-d1ff-4c42-847a-9538961b1a6a.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.903658] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquired lock "[datastore1] devstack-image-cache_base/7e436a80-d1ff-4c42-847a-9538961b1a6a/7e436a80-d1ff-4c42-847a-9538961b1a6a.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.903834] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 901.904102] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-da872866-3e20-4c7e-b0fb-03edc02e63aa {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.906928] env[62965]: DEBUG oslo_concurrency.lockutils [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Releasing lock "refresh_cache-892b1302-f536-4dcd-9227-a545ae1ac731" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.906928] env[62965]: DEBUG nova.compute.manager [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Instance network_info: |[{"id": "6205e686-0f1c-46bf-a1e7-88f0fd8ee5e5", "address": "fa:16:3e:99:64:c5", "network": {"id": "a0a5e605-d580-4b7b-b3c0-5c7395bd5a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1679618017-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ed71548c085499981fa1b102d660368", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6205e686-0f", "ovs_interfaceid": "6205e686-0f1c-46bf-a1e7-88f0fd8ee5e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 901.907082] env[62965]: DEBUG oslo_concurrency.lockutils [req-d2d5db2c-94f0-48fa-a429-de5175cdf0e6 req-b50756f7-5b12-4311-b66d-47fd1a91c32c service nova] Acquired lock "refresh_cache-892b1302-f536-4dcd-9227-a545ae1ac731" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.907082] env[62965]: DEBUG nova.network.neutron [req-d2d5db2c-94f0-48fa-a429-de5175cdf0e6 req-b50756f7-5b12-4311-b66d-47fd1a91c32c service nova] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Refreshing network info cache for port 6205e686-0f1c-46bf-a1e7-88f0fd8ee5e5 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 901.909207] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:99:64:c5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0cd5d325-3053-407e-a4ee-f627e82a23f9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6205e686-0f1c-46bf-a1e7-88f0fd8ee5e5', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 901.916445] env[62965]: DEBUG oslo.service.loopingcall [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 901.919492] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 901.920025] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3a73e3f1-4965-4c3a-a336-ff0e736afe41 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.937039] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 901.937255] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 901.938022] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a53626a-f36c-460a-b4b2-f6d5b1196194 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.941485] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 901.941485] env[62965]: value = "task-1389967" [ 901.941485] env[62965]: _type = "Task" [ 901.941485] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.945953] env[62965]: DEBUG oslo_vmware.api [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 901.945953] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5285fc54-300e-8b3e-057d-23ba1bd5d28d" [ 901.945953] env[62965]: _type = "Task" [ 901.945953] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.952168] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389967, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.958143] env[62965]: DEBUG oslo_vmware.api [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5285fc54-300e-8b3e-057d-23ba1bd5d28d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.029838] env[62965]: INFO nova.compute.manager [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Took 15.40 seconds to build instance. [ 902.184290] env[62965]: DEBUG nova.compute.manager [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 902.217092] env[62965]: DEBUG nova.network.neutron [req-d2d5db2c-94f0-48fa-a429-de5175cdf0e6 req-b50756f7-5b12-4311-b66d-47fd1a91c32c service nova] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Updated VIF entry in instance network info cache for port 6205e686-0f1c-46bf-a1e7-88f0fd8ee5e5. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 902.217486] env[62965]: DEBUG nova.network.neutron [req-d2d5db2c-94f0-48fa-a429-de5175cdf0e6 req-b50756f7-5b12-4311-b66d-47fd1a91c32c service nova] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Updating instance_info_cache with network_info: [{"id": "6205e686-0f1c-46bf-a1e7-88f0fd8ee5e5", "address": "fa:16:3e:99:64:c5", "network": {"id": "a0a5e605-d580-4b7b-b3c0-5c7395bd5a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1679618017-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ed71548c085499981fa1b102d660368", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6205e686-0f", "ovs_interfaceid": "6205e686-0f1c-46bf-a1e7-88f0fd8ee5e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.234922] env[62965]: DEBUG nova.virt.hardware [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 902.235208] env[62965]: DEBUG nova.virt.hardware [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 902.235383] env[62965]: DEBUG nova.virt.hardware [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 902.235571] env[62965]: DEBUG nova.virt.hardware [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 902.235724] env[62965]: DEBUG nova.virt.hardware [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 902.235879] env[62965]: DEBUG nova.virt.hardware [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 902.236119] env[62965]: DEBUG nova.virt.hardware [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 902.236320] env[62965]: DEBUG nova.virt.hardware [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 902.237493] env[62965]: DEBUG nova.virt.hardware [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 902.237493] env[62965]: DEBUG nova.virt.hardware [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 902.237493] env[62965]: DEBUG nova.virt.hardware [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 902.238352] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d43d11b2-0f42-4db9-92e3-983ede2afe25 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.252028] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83c17cd3-0065-4bec-a783-dd72dd67388d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.398133] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c9895e6-9ec6-4bc3-8a85-a4cdc926665f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Acquiring lock "64cf898c-2c97-4daa-80c8-0a1c252cbcf2" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.398364] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c9895e6-9ec6-4bc3-8a85-a4cdc926665f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Lock "64cf898c-2c97-4daa-80c8-0a1c252cbcf2" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.458139] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389967, 'name': CreateVM_Task} progress is 99%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.464379] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Preparing fetch location {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 902.464638] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Fetch image to [datastore1] OSTACK_IMG_9310edad-f829-4b40-8be5-438377eac181/OSTACK_IMG_9310edad-f829-4b40-8be5-438377eac181.vmdk {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 902.464821] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Downloading stream optimized image 7e436a80-d1ff-4c42-847a-9538961b1a6a to [datastore1] OSTACK_IMG_9310edad-f829-4b40-8be5-438377eac181/OSTACK_IMG_9310edad-f829-4b40-8be5-438377eac181.vmdk on the data store datastore1 as vApp {{(pid=62965) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 902.464993] env[62965]: DEBUG nova.virt.vmwareapi.images [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Downloading image file data 7e436a80-d1ff-4c42-847a-9538961b1a6a to the ESX as VM named 'OSTACK_IMG_9310edad-f829-4b40-8be5-438377eac181' {{(pid=62965) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 902.524496] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-657803bf-fb99-4607-9266-2895f407de34 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.533998] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b4e51ad7-46e9-4565-a74b-de48a48bb1bd tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Lock "fa511b61-887e-46a1-aaf6-586d716c0fb5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.924s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.535389] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-156fb562-d80e-4af9-ace7-1a74d598f98e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.568180] env[62965]: DEBUG oslo_vmware.rw_handles [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 902.568180] env[62965]: value = "resgroup-9" [ 902.568180] env[62965]: _type = "ResourcePool" [ 902.568180] env[62965]: }. {{(pid=62965) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 902.569096] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01525659-e6dc-40f7-b16c-22cd6a86fdb7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.572021] env[62965]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-d1939fd2-044d-4e0d-b2f0-c75155833200 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.589358] env[62965]: DEBUG nova.network.neutron [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Successfully updated port: 838a71f4-1d2e-4ffd-8ba7-a09d951c4a66 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 902.595772] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6fe7e3f-8753-4a2e-bcb8-3553d55fd2e6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.601957] env[62965]: DEBUG oslo_vmware.rw_handles [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lease: (returnval){ [ 902.601957] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5265c4f0-3740-4a5c-4484-a3aa7a42cfde" [ 902.601957] env[62965]: _type = "HttpNfcLease" [ 902.601957] env[62965]: } obtained for vApp import into resource pool (val){ [ 902.601957] env[62965]: value = "resgroup-9" [ 902.601957] env[62965]: _type = "ResourcePool" [ 902.601957] env[62965]: }. {{(pid=62965) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 902.602258] env[62965]: DEBUG oslo_vmware.api [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the lease: (returnval){ [ 902.602258] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5265c4f0-3740-4a5c-4484-a3aa7a42cfde" [ 902.602258] env[62965]: _type = "HttpNfcLease" [ 902.602258] env[62965]: } to be ready. {{(pid=62965) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 902.613967] env[62965]: DEBUG nova.compute.provider_tree [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 902.619691] env[62965]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 902.619691] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5265c4f0-3740-4a5c-4484-a3aa7a42cfde" [ 902.619691] env[62965]: _type = "HttpNfcLease" [ 902.619691] env[62965]: } is initializing. {{(pid=62965) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 902.654844] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ba31b606-70ea-44b4-80d5-2d58061ebaa5 tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Acquiring lock "fa511b61-887e-46a1-aaf6-586d716c0fb5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.655130] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ba31b606-70ea-44b4-80d5-2d58061ebaa5 tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Lock "fa511b61-887e-46a1-aaf6-586d716c0fb5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.655348] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ba31b606-70ea-44b4-80d5-2d58061ebaa5 tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Acquiring lock "fa511b61-887e-46a1-aaf6-586d716c0fb5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.655529] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ba31b606-70ea-44b4-80d5-2d58061ebaa5 tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Lock "fa511b61-887e-46a1-aaf6-586d716c0fb5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.655698] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ba31b606-70ea-44b4-80d5-2d58061ebaa5 tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Lock "fa511b61-887e-46a1-aaf6-586d716c0fb5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.658270] env[62965]: INFO nova.compute.manager [None req-ba31b606-70ea-44b4-80d5-2d58061ebaa5 tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Terminating instance [ 902.721008] env[62965]: DEBUG oslo_concurrency.lockutils [req-d2d5db2c-94f0-48fa-a429-de5175cdf0e6 req-b50756f7-5b12-4311-b66d-47fd1a91c32c service nova] Releasing lock "refresh_cache-892b1302-f536-4dcd-9227-a545ae1ac731" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.904642] env[62965]: DEBUG nova.compute.utils [None req-7c9895e6-9ec6-4bc3-8a85-a4cdc926665f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 902.960162] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389967, 'name': CreateVM_Task} progress is 99%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.093908] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquiring lock "refresh_cache-eed55d8e-21dc-47d4-a689-415ef6a3891d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.094088] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquired lock "refresh_cache-eed55d8e-21dc-47d4-a689-415ef6a3891d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.094244] env[62965]: DEBUG nova.network.neutron [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 903.112955] env[62965]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 903.112955] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5265c4f0-3740-4a5c-4484-a3aa7a42cfde" [ 903.112955] env[62965]: _type = "HttpNfcLease" [ 903.112955] env[62965]: } is initializing. {{(pid=62965) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 903.116796] env[62965]: DEBUG nova.scheduler.client.report [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 903.161464] env[62965]: DEBUG nova.compute.manager [None req-ba31b606-70ea-44b4-80d5-2d58061ebaa5 tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 903.162459] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ba31b606-70ea-44b4-80d5-2d58061ebaa5 tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 903.163374] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e392cafa-5de4-4c6f-a946-3e519d97f156 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.170699] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba31b606-70ea-44b4-80d5-2d58061ebaa5 tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 903.170967] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8f4c6d47-b6af-457d-91e1-9d1fb1146fa0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.177656] env[62965]: DEBUG oslo_vmware.api [None req-ba31b606-70ea-44b4-80d5-2d58061ebaa5 tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Waiting for the task: (returnval){ [ 903.177656] env[62965]: value = "task-1389969" [ 903.177656] env[62965]: _type = "Task" [ 903.177656] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.185049] env[62965]: DEBUG oslo_vmware.api [None req-ba31b606-70ea-44b4-80d5-2d58061ebaa5 tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Task: {'id': task-1389969, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.395075] env[62965]: DEBUG nova.compute.manager [req-838e6122-89b3-4e6a-9372-fa5bdbd7e24a req-15f450ec-9925-4d0d-925a-980b884d6061 service nova] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Received event network-vif-plugged-838a71f4-1d2e-4ffd-8ba7-a09d951c4a66 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 903.395424] env[62965]: DEBUG oslo_concurrency.lockutils [req-838e6122-89b3-4e6a-9372-fa5bdbd7e24a req-15f450ec-9925-4d0d-925a-980b884d6061 service nova] Acquiring lock "eed55d8e-21dc-47d4-a689-415ef6a3891d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.395424] env[62965]: DEBUG oslo_concurrency.lockutils [req-838e6122-89b3-4e6a-9372-fa5bdbd7e24a req-15f450ec-9925-4d0d-925a-980b884d6061 service nova] Lock "eed55d8e-21dc-47d4-a689-415ef6a3891d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.395613] env[62965]: DEBUG oslo_concurrency.lockutils [req-838e6122-89b3-4e6a-9372-fa5bdbd7e24a req-15f450ec-9925-4d0d-925a-980b884d6061 service nova] Lock "eed55d8e-21dc-47d4-a689-415ef6a3891d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.395762] env[62965]: DEBUG nova.compute.manager [req-838e6122-89b3-4e6a-9372-fa5bdbd7e24a req-15f450ec-9925-4d0d-925a-980b884d6061 service nova] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] No waiting events found dispatching network-vif-plugged-838a71f4-1d2e-4ffd-8ba7-a09d951c4a66 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 903.395932] env[62965]: WARNING nova.compute.manager [req-838e6122-89b3-4e6a-9372-fa5bdbd7e24a req-15f450ec-9925-4d0d-925a-980b884d6061 service nova] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Received unexpected event network-vif-plugged-838a71f4-1d2e-4ffd-8ba7-a09d951c4a66 for instance with vm_state building and task_state spawning. [ 903.396104] env[62965]: DEBUG nova.compute.manager [req-838e6122-89b3-4e6a-9372-fa5bdbd7e24a req-15f450ec-9925-4d0d-925a-980b884d6061 service nova] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Received event network-changed-838a71f4-1d2e-4ffd-8ba7-a09d951c4a66 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 903.396260] env[62965]: DEBUG nova.compute.manager [req-838e6122-89b3-4e6a-9372-fa5bdbd7e24a req-15f450ec-9925-4d0d-925a-980b884d6061 service nova] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Refreshing instance network info cache due to event network-changed-838a71f4-1d2e-4ffd-8ba7-a09d951c4a66. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 903.396427] env[62965]: DEBUG oslo_concurrency.lockutils [req-838e6122-89b3-4e6a-9372-fa5bdbd7e24a req-15f450ec-9925-4d0d-925a-980b884d6061 service nova] Acquiring lock "refresh_cache-eed55d8e-21dc-47d4-a689-415ef6a3891d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.408033] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c9895e6-9ec6-4bc3-8a85-a4cdc926665f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Lock "64cf898c-2c97-4daa-80c8-0a1c252cbcf2" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.455837] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389967, 'name': CreateVM_Task} progress is 99%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.612070] env[62965]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 903.612070] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5265c4f0-3740-4a5c-4484-a3aa7a42cfde" [ 903.612070] env[62965]: _type = "HttpNfcLease" [ 903.612070] env[62965]: } is initializing. {{(pid=62965) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 903.621076] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.443s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.621648] env[62965]: DEBUG nova.compute.manager [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 903.624318] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.955s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.625698] env[62965]: INFO nova.compute.claims [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 903.628692] env[62965]: DEBUG nova.network.neutron [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 903.686607] env[62965]: DEBUG oslo_vmware.api [None req-ba31b606-70ea-44b4-80d5-2d58061ebaa5 tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Task: {'id': task-1389969, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.765618] env[62965]: DEBUG nova.network.neutron [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Updating instance_info_cache with network_info: [{"id": "838a71f4-1d2e-4ffd-8ba7-a09d951c4a66", "address": "fa:16:3e:f7:d0:27", "network": {"id": "e9a6ebdd-0479-45af-b947-5d35ae182c87", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-385052062-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89fd5275b2064288b44e79365f6f0271", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ba07329-1d3e-4ba8-8774-d029262318c4", "external-id": "nsx-vlan-transportzone-534", "segmentation_id": 534, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap838a71f4-1d", "ovs_interfaceid": "838a71f4-1d2e-4ffd-8ba7-a09d951c4a66", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.957542] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389967, 'name': CreateVM_Task, 'duration_secs': 1.581906} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.957679] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 903.958337] env[62965]: DEBUG oslo_concurrency.lockutils [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.958564] env[62965]: DEBUG oslo_concurrency.lockutils [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.958855] env[62965]: DEBUG oslo_concurrency.lockutils [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 903.959203] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a01bcb2-3b1b-485b-bf73-78c255a4634e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.963955] env[62965]: DEBUG oslo_vmware.api [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 903.963955] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52213f57-6cf8-3b49-d431-d32a8b40185d" [ 903.963955] env[62965]: _type = "Task" [ 903.963955] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.975505] env[62965]: DEBUG oslo_vmware.api [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52213f57-6cf8-3b49-d431-d32a8b40185d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.113371] env[62965]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 904.113371] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5265c4f0-3740-4a5c-4484-a3aa7a42cfde" [ 904.113371] env[62965]: _type = "HttpNfcLease" [ 904.113371] env[62965]: } is initializing. {{(pid=62965) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 904.130592] env[62965]: DEBUG nova.compute.utils [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 904.132221] env[62965]: DEBUG nova.compute.manager [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 904.132443] env[62965]: DEBUG nova.network.neutron [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 904.185405] env[62965]: DEBUG nova.policy [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '736eaca6d3584f74856c61ff0e6ce03e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd3edf0643c9d40e9bb568742b7a7a508', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 904.191870] env[62965]: DEBUG oslo_vmware.api [None req-ba31b606-70ea-44b4-80d5-2d58061ebaa5 tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Task: {'id': task-1389969, 'name': PowerOffVM_Task, 'duration_secs': 0.629697} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.192564] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba31b606-70ea-44b4-80d5-2d58061ebaa5 tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 904.192564] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ba31b606-70ea-44b4-80d5-2d58061ebaa5 tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 904.193362] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-58c21fe7-05df-4212-ab5d-31e94c5fdcbb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.257285] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ba31b606-70ea-44b4-80d5-2d58061ebaa5 tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 904.257529] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ba31b606-70ea-44b4-80d5-2d58061ebaa5 tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Deleting contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 904.257843] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba31b606-70ea-44b4-80d5-2d58061ebaa5 tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Deleting the datastore file [datastore2] fa511b61-887e-46a1-aaf6-586d716c0fb5 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 904.258167] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bb37c5b1-b7f1-46f5-bfd7-cc79018c9427 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.267600] env[62965]: DEBUG oslo_vmware.api [None req-ba31b606-70ea-44b4-80d5-2d58061ebaa5 tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Waiting for the task: (returnval){ [ 904.267600] env[62965]: value = "task-1389971" [ 904.267600] env[62965]: _type = "Task" [ 904.267600] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.268065] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Releasing lock "refresh_cache-eed55d8e-21dc-47d4-a689-415ef6a3891d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.268340] env[62965]: DEBUG nova.compute.manager [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Instance network_info: |[{"id": "838a71f4-1d2e-4ffd-8ba7-a09d951c4a66", "address": "fa:16:3e:f7:d0:27", "network": {"id": "e9a6ebdd-0479-45af-b947-5d35ae182c87", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-385052062-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89fd5275b2064288b44e79365f6f0271", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ba07329-1d3e-4ba8-8774-d029262318c4", "external-id": "nsx-vlan-transportzone-534", "segmentation_id": 534, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap838a71f4-1d", "ovs_interfaceid": "838a71f4-1d2e-4ffd-8ba7-a09d951c4a66", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 904.271467] env[62965]: DEBUG oslo_concurrency.lockutils [req-838e6122-89b3-4e6a-9372-fa5bdbd7e24a req-15f450ec-9925-4d0d-925a-980b884d6061 service nova] Acquired lock "refresh_cache-eed55d8e-21dc-47d4-a689-415ef6a3891d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.271698] env[62965]: DEBUG nova.network.neutron [req-838e6122-89b3-4e6a-9372-fa5bdbd7e24a req-15f450ec-9925-4d0d-925a-980b884d6061 service nova] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Refreshing network info cache for port 838a71f4-1d2e-4ffd-8ba7-a09d951c4a66 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 904.272935] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f7:d0:27', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5ba07329-1d3e-4ba8-8774-d029262318c4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '838a71f4-1d2e-4ffd-8ba7-a09d951c4a66', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 904.280805] env[62965]: DEBUG oslo.service.loopingcall [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 904.283592] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 904.284717] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c3aa5ade-f3b5-4f36-895a-348ada54877c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.303014] env[62965]: DEBUG oslo_vmware.api [None req-ba31b606-70ea-44b4-80d5-2d58061ebaa5 tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Task: {'id': task-1389971, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.307678] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 904.307678] env[62965]: value = "task-1389972" [ 904.307678] env[62965]: _type = "Task" [ 904.307678] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.318373] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389972, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.474879] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c9895e6-9ec6-4bc3-8a85-a4cdc926665f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Acquiring lock "64cf898c-2c97-4daa-80c8-0a1c252cbcf2" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.475214] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c9895e6-9ec6-4bc3-8a85-a4cdc926665f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Lock "64cf898c-2c97-4daa-80c8-0a1c252cbcf2" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.475491] env[62965]: INFO nova.compute.manager [None req-7c9895e6-9ec6-4bc3-8a85-a4cdc926665f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Attaching volume eeae4205-49ab-42e1-9d79-4b17b3886993 to /dev/sdb [ 904.477288] env[62965]: DEBUG oslo_vmware.api [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52213f57-6cf8-3b49-d431-d32a8b40185d, 'name': SearchDatastore_Task, 'duration_secs': 0.019307} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.478075] env[62965]: DEBUG oslo_concurrency.lockutils [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.478375] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 904.478641] env[62965]: DEBUG oslo_concurrency.lockutils [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.478939] env[62965]: DEBUG oslo_concurrency.lockutils [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.479219] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 904.479712] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b4414462-e7d2-45b8-8608-c1f2de4f4511 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.491834] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 904.492085] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 904.492795] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d8a86265-2713-4fe9-8d74-4d53da5b166e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.500879] env[62965]: DEBUG oslo_vmware.api [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 904.500879] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5210d32b-9a5b-964d-82eb-8f508da979d9" [ 904.500879] env[62965]: _type = "Task" [ 904.500879] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.512543] env[62965]: DEBUG oslo_vmware.api [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5210d32b-9a5b-964d-82eb-8f508da979d9, 'name': SearchDatastore_Task, 'duration_secs': 0.008761} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.513435] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69d78b26-0f51-4f17-b5bb-12a8aac7d8bb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.519479] env[62965]: DEBUG oslo_vmware.api [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 904.519479] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]526c10f1-6c76-b7c7-6ef5-f799e377118b" [ 904.519479] env[62965]: _type = "Task" [ 904.519479] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.521768] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd376316-9420-4e72-8091-b05f0fe592d3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.534924] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e51df987-eecc-49a9-8aff-f9ae40e2750c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.537423] env[62965]: DEBUG oslo_vmware.api [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]526c10f1-6c76-b7c7-6ef5-f799e377118b, 'name': SearchDatastore_Task, 'duration_secs': 0.007778} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.537683] env[62965]: DEBUG oslo_concurrency.lockutils [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.537973] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] 892b1302-f536-4dcd-9227-a545ae1ac731/892b1302-f536-4dcd-9227-a545ae1ac731.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 904.538591] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b2718934-0e94-4c76-a6ab-bca0973b7475 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.545079] env[62965]: DEBUG oslo_vmware.api [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 904.545079] env[62965]: value = "task-1389973" [ 904.545079] env[62965]: _type = "Task" [ 904.545079] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.548573] env[62965]: DEBUG nova.virt.block_device [None req-7c9895e6-9ec6-4bc3-8a85-a4cdc926665f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Updating existing volume attachment record: bbe18756-0a53-4121-80eb-7292cdbedbd1 {{(pid=62965) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 904.558495] env[62965]: DEBUG oslo_vmware.api [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1389973, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.613348] env[62965]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 904.613348] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5265c4f0-3740-4a5c-4484-a3aa7a42cfde" [ 904.613348] env[62965]: _type = "HttpNfcLease" [ 904.613348] env[62965]: } is initializing. {{(pid=62965) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 904.638232] env[62965]: DEBUG nova.compute.manager [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 904.757328] env[62965]: DEBUG nova.network.neutron [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Successfully created port: fe1bf4cc-7773-462b-b8b5-1849c661d910 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 904.778772] env[62965]: DEBUG oslo_vmware.api [None req-ba31b606-70ea-44b4-80d5-2d58061ebaa5 tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Task: {'id': task-1389971, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133823} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.782614] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba31b606-70ea-44b4-80d5-2d58061ebaa5 tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 904.782912] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ba31b606-70ea-44b4-80d5-2d58061ebaa5 tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Deleted contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 904.783201] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ba31b606-70ea-44b4-80d5-2d58061ebaa5 tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 904.783431] env[62965]: INFO nova.compute.manager [None req-ba31b606-70ea-44b4-80d5-2d58061ebaa5 tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Took 1.62 seconds to destroy the instance on the hypervisor. [ 904.783727] env[62965]: DEBUG oslo.service.loopingcall [None req-ba31b606-70ea-44b4-80d5-2d58061ebaa5 tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 904.784214] env[62965]: DEBUG nova.compute.manager [-] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 904.784492] env[62965]: DEBUG nova.network.neutron [-] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 904.820970] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389972, 'name': CreateVM_Task, 'duration_secs': 0.317377} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.821326] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 904.822248] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.822525] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.823028] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 904.823354] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b40a4a0-51a9-435f-8c67-e31519432455 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.826767] env[62965]: DEBUG nova.network.neutron [req-838e6122-89b3-4e6a-9372-fa5bdbd7e24a req-15f450ec-9925-4d0d-925a-980b884d6061 service nova] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Updated VIF entry in instance network info cache for port 838a71f4-1d2e-4ffd-8ba7-a09d951c4a66. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 904.827205] env[62965]: DEBUG nova.network.neutron [req-838e6122-89b3-4e6a-9372-fa5bdbd7e24a req-15f450ec-9925-4d0d-925a-980b884d6061 service nova] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Updating instance_info_cache with network_info: [{"id": "838a71f4-1d2e-4ffd-8ba7-a09d951c4a66", "address": "fa:16:3e:f7:d0:27", "network": {"id": "e9a6ebdd-0479-45af-b947-5d35ae182c87", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-385052062-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89fd5275b2064288b44e79365f6f0271", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ba07329-1d3e-4ba8-8774-d029262318c4", "external-id": "nsx-vlan-transportzone-534", "segmentation_id": 534, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap838a71f4-1d", "ovs_interfaceid": "838a71f4-1d2e-4ffd-8ba7-a09d951c4a66", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.834763] env[62965]: DEBUG oslo_vmware.api [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 904.834763] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]522967da-6baf-085a-53ce-3c06370e005e" [ 904.834763] env[62965]: _type = "Task" [ 904.834763] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.852339] env[62965]: DEBUG oslo_vmware.api [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]522967da-6baf-085a-53ce-3c06370e005e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.998416] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-799a5391-7239-4f9d-9cd5-4a884e4f1448 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.010138] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-560dcb08-7a30-4470-8845-cb001e907de0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.046736] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4ccff5c-a136-4b91-8d15-21df8747d663 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.063127] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac64ba2e-25d6-422d-9522-b736ce9d07f2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.067597] env[62965]: DEBUG oslo_vmware.api [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1389973, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.079096] env[62965]: DEBUG nova.compute.provider_tree [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 905.116576] env[62965]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 905.116576] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5265c4f0-3740-4a5c-4484-a3aa7a42cfde" [ 905.116576] env[62965]: _type = "HttpNfcLease" [ 905.116576] env[62965]: } is initializing. {{(pid=62965) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 905.330567] env[62965]: DEBUG oslo_concurrency.lockutils [req-838e6122-89b3-4e6a-9372-fa5bdbd7e24a req-15f450ec-9925-4d0d-925a-980b884d6061 service nova] Releasing lock "refresh_cache-eed55d8e-21dc-47d4-a689-415ef6a3891d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.346880] env[62965]: DEBUG oslo_vmware.api [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]522967da-6baf-085a-53ce-3c06370e005e, 'name': SearchDatastore_Task, 'duration_secs': 0.073669} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.347023] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.347241] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 905.347472] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.347620] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.347813] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 905.348106] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e091c1ae-a0b4-4b18-b476-cccc69f2c14d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.356485] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 905.356780] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 905.357629] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7e476584-c388-4ef7-ac7e-680855fb82e8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.362861] env[62965]: DEBUG oslo_vmware.api [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 905.362861] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]520d2328-fad7-e2d6-fc71-ab9756435dec" [ 905.362861] env[62965]: _type = "Task" [ 905.362861] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.370963] env[62965]: DEBUG oslo_vmware.api [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]520d2328-fad7-e2d6-fc71-ab9756435dec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.426831] env[62965]: DEBUG nova.compute.manager [req-74923c68-5d0a-4d7d-8427-63623239bce8 req-3cf2cbe5-ee1b-42d2-a1e9-daa4c4804db5 service nova] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Received event network-vif-deleted-34d9279a-1a85-4e01-8a4f-b07cb77ec0e6 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 905.427009] env[62965]: INFO nova.compute.manager [req-74923c68-5d0a-4d7d-8427-63623239bce8 req-3cf2cbe5-ee1b-42d2-a1e9-daa4c4804db5 service nova] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Neutron deleted interface 34d9279a-1a85-4e01-8a4f-b07cb77ec0e6; detaching it from the instance and deleting it from the info cache [ 905.427180] env[62965]: DEBUG nova.network.neutron [req-74923c68-5d0a-4d7d-8427-63623239bce8 req-3cf2cbe5-ee1b-42d2-a1e9-daa4c4804db5 service nova] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.559409] env[62965]: DEBUG oslo_vmware.api [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1389973, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.637343} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.560084] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] 892b1302-f536-4dcd-9227-a545ae1ac731/892b1302-f536-4dcd-9227-a545ae1ac731.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 905.560311] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 905.560575] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-521fa670-d14e-44b0-a404-15e2ec946bcd {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.566710] env[62965]: DEBUG oslo_vmware.api [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 905.566710] env[62965]: value = "task-1389977" [ 905.566710] env[62965]: _type = "Task" [ 905.566710] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.574351] env[62965]: DEBUG oslo_vmware.api [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1389977, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.600525] env[62965]: ERROR nova.scheduler.client.report [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [req-1766b238-1326-47fa-85c4-51fffe9e802d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-1766b238-1326-47fa-85c4-51fffe9e802d"}]} [ 905.614726] env[62965]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 905.614726] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5265c4f0-3740-4a5c-4484-a3aa7a42cfde" [ 905.614726] env[62965]: _type = "HttpNfcLease" [ 905.614726] env[62965]: } is ready. {{(pid=62965) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 905.615043] env[62965]: DEBUG oslo_vmware.rw_handles [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 905.615043] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5265c4f0-3740-4a5c-4484-a3aa7a42cfde" [ 905.615043] env[62965]: _type = "HttpNfcLease" [ 905.615043] env[62965]: }. {{(pid=62965) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 905.615731] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a80c043-298b-4747-a61a-bd2439ecc5c5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.618723] env[62965]: DEBUG nova.scheduler.client.report [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Refreshing inventories for resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 905.625156] env[62965]: DEBUG oslo_vmware.rw_handles [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52843a9b-813c-e508-1bd7-e28e4eb3e905/disk-0.vmdk from lease info. {{(pid=62965) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 905.625493] env[62965]: DEBUG oslo_vmware.rw_handles [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52843a9b-813c-e508-1bd7-e28e4eb3e905/disk-0.vmdk. {{(pid=62965) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 905.681893] env[62965]: DEBUG nova.compute.manager [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 905.684669] env[62965]: DEBUG nova.scheduler.client.report [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Updating ProviderTree inventory for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 905.684875] env[62965]: DEBUG nova.compute.provider_tree [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 905.694832] env[62965]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-001c408e-a077-4b0b-a93a-8f79d57f11ee {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.700577] env[62965]: DEBUG nova.scheduler.client.report [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Refreshing aggregate associations for resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8, aggregates: None {{(pid=62965) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 905.712176] env[62965]: DEBUG nova.virt.hardware [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 905.712509] env[62965]: DEBUG nova.virt.hardware [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 905.712676] env[62965]: DEBUG nova.virt.hardware [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 905.712864] env[62965]: DEBUG nova.virt.hardware [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 905.713023] env[62965]: DEBUG nova.virt.hardware [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 905.713241] env[62965]: DEBUG nova.virt.hardware [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 905.713476] env[62965]: DEBUG nova.virt.hardware [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 905.713655] env[62965]: DEBUG nova.virt.hardware [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 905.713836] env[62965]: DEBUG nova.virt.hardware [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 905.714024] env[62965]: DEBUG nova.virt.hardware [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 905.714223] env[62965]: DEBUG nova.virt.hardware [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 905.715302] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-726c5b4a-cc40-4e7a-8e23-4dcd273050c9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.722920] env[62965]: DEBUG nova.scheduler.client.report [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Refreshing trait associations for resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64 {{(pid=62965) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 905.728255] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1affc04-b4a5-405d-acef-a4298227ab8d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.743269] env[62965]: DEBUG nova.network.neutron [-] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.875756] env[62965]: DEBUG oslo_vmware.api [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]520d2328-fad7-e2d6-fc71-ab9756435dec, 'name': SearchDatastore_Task, 'duration_secs': 0.015157} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.876626] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-04a48e2e-4ecb-4596-9e6d-58e7e9c65aec {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.882118] env[62965]: DEBUG oslo_vmware.api [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 905.882118] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52270883-675b-536a-b280-47b0cb664226" [ 905.882118] env[62965]: _type = "Task" [ 905.882118] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.892579] env[62965]: DEBUG oslo_vmware.api [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52270883-675b-536a-b280-47b0cb664226, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.935023] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7bead058-19ae-4940-8922-f61fc0c38680 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.954849] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd5571cd-0c04-421f-b205-dbfe0d0f5bbc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.006884] env[62965]: DEBUG nova.compute.manager [req-74923c68-5d0a-4d7d-8427-63623239bce8 req-3cf2cbe5-ee1b-42d2-a1e9-daa4c4804db5 service nova] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Detach interface failed, port_id=34d9279a-1a85-4e01-8a4f-b07cb77ec0e6, reason: Instance fa511b61-887e-46a1-aaf6-586d716c0fb5 could not be found. {{(pid=62965) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11294}} [ 906.078365] env[62965]: DEBUG oslo_vmware.api [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1389977, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.204884} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.082211] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 906.084836] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b51ce9bb-7f48-411d-bbff-98dd40d23911 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.108941] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] 892b1302-f536-4dcd-9227-a545ae1ac731/892b1302-f536-4dcd-9227-a545ae1ac731.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 906.117181] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2d75dcb7-47c1-4ad1-b614-fc12350309e3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.143294] env[62965]: DEBUG oslo_vmware.api [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 906.143294] env[62965]: value = "task-1389978" [ 906.143294] env[62965]: _type = "Task" [ 906.143294] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.150461] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0c3fe85-e205-428a-947a-189815a5b4db {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.163919] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed87aca7-726f-4afe-8741-b667075d809a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.167390] env[62965]: DEBUG oslo_vmware.api [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1389978, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.206224] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d234a5d4-1ead-4b42-9314-3ce9b402c227 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.215209] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aefb55e2-67cc-471e-af9c-157c9079bf1d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.235555] env[62965]: DEBUG nova.compute.provider_tree [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 906.247194] env[62965]: INFO nova.compute.manager [-] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Took 1.46 seconds to deallocate network for instance. [ 906.392981] env[62965]: DEBUG oslo_vmware.api [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52270883-675b-536a-b280-47b0cb664226, 'name': SearchDatastore_Task, 'duration_secs': 0.010117} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.396583] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.396920] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] eed55d8e-21dc-47d4-a689-415ef6a3891d/eed55d8e-21dc-47d4-a689-415ef6a3891d.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 906.397237] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e7eab791-8017-4231-a693-6e321df6cc8e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.406853] env[62965]: DEBUG oslo_vmware.api [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 906.406853] env[62965]: value = "task-1389979" [ 906.406853] env[62965]: _type = "Task" [ 906.406853] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.418112] env[62965]: DEBUG oslo_vmware.api [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1389979, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.459930] env[62965]: DEBUG oslo_vmware.rw_handles [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Completed reading data from the image iterator. {{(pid=62965) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 906.460390] env[62965]: DEBUG oslo_vmware.rw_handles [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52843a9b-813c-e508-1bd7-e28e4eb3e905/disk-0.vmdk. {{(pid=62965) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 906.461718] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16a72c50-2b1d-4b72-8f40-66a948c54c49 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.468956] env[62965]: DEBUG oslo_vmware.rw_handles [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52843a9b-813c-e508-1bd7-e28e4eb3e905/disk-0.vmdk is in state: ready. {{(pid=62965) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 906.469089] env[62965]: DEBUG oslo_vmware.rw_handles [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52843a9b-813c-e508-1bd7-e28e4eb3e905/disk-0.vmdk. {{(pid=62965) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 906.469366] env[62965]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-d5f116a4-83c7-4706-b5e2-cfbbfcfa83e7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.547424] env[62965]: DEBUG nova.network.neutron [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Successfully updated port: fe1bf4cc-7773-462b-b8b5-1849c661d910 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 906.651342] env[62965]: DEBUG oslo_vmware.rw_handles [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52843a9b-813c-e508-1bd7-e28e4eb3e905/disk-0.vmdk. {{(pid=62965) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 906.651613] env[62965]: INFO nova.virt.vmwareapi.images [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Downloaded image file data 7e436a80-d1ff-4c42-847a-9538961b1a6a [ 906.652957] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-712086e1-ee80-4926-9d2d-928416b1f6b3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.659597] env[62965]: DEBUG oslo_vmware.api [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1389978, 'name': ReconfigVM_Task, 'duration_secs': 0.404315} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.660285] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Reconfigured VM instance instance-0000004f to attach disk [datastore1] 892b1302-f536-4dcd-9227-a545ae1ac731/892b1302-f536-4dcd-9227-a545ae1ac731.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 906.660942] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f2d49bdf-2ac8-4f00-8807-9ab2930dc1be {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.673559] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bc1211a9-ff8d-408d-b381-0207fa6ed41e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.677786] env[62965]: DEBUG oslo_vmware.api [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 906.677786] env[62965]: value = "task-1389980" [ 906.677786] env[62965]: _type = "Task" [ 906.677786] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.686251] env[62965]: DEBUG oslo_vmware.api [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1389980, 'name': Rename_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.712732] env[62965]: INFO nova.virt.vmwareapi.images [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] The imported VM was unregistered [ 906.715232] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Caching image {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 906.715470] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Creating directory with path [datastore1] devstack-image-cache_base/7e436a80-d1ff-4c42-847a-9538961b1a6a {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 906.715805] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-20ec2524-d7d4-4f7e-8cdc-d09a13afd56d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.756584] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Created directory with path [datastore1] devstack-image-cache_base/7e436a80-d1ff-4c42-847a-9538961b1a6a {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 906.756871] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_9310edad-f829-4b40-8be5-438377eac181/OSTACK_IMG_9310edad-f829-4b40-8be5-438377eac181.vmdk to [datastore1] devstack-image-cache_base/7e436a80-d1ff-4c42-847a-9538961b1a6a/7e436a80-d1ff-4c42-847a-9538961b1a6a.vmdk. {{(pid=62965) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 906.758034] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ba31b606-70ea-44b4-80d5-2d58061ebaa5 tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.758210] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-4d4f0abc-0353-49d7-b64c-b8ed2d90fd65 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.765866] env[62965]: DEBUG oslo_vmware.api [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 906.765866] env[62965]: value = "task-1389982" [ 906.765866] env[62965]: _type = "Task" [ 906.765866] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.774859] env[62965]: DEBUG oslo_vmware.api [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389982, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.778904] env[62965]: DEBUG nova.scheduler.client.report [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Updated inventory for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with generation 91 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 906.779206] env[62965]: DEBUG nova.compute.provider_tree [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Updating resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 generation from 91 to 92 during operation: update_inventory {{(pid=62965) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 906.779399] env[62965]: DEBUG nova.compute.provider_tree [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 906.919871] env[62965]: DEBUG oslo_vmware.api [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1389979, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.051389] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Acquiring lock "refresh_cache-e0c4e188-a041-4aa7-95de-6e8718b17ade" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.051477] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Acquired lock "refresh_cache-e0c4e188-a041-4aa7-95de-6e8718b17ade" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.051603] env[62965]: DEBUG nova.network.neutron [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 907.187728] env[62965]: DEBUG oslo_vmware.api [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1389980, 'name': Rename_Task, 'duration_secs': 0.215697} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.188045] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 907.188301] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-57994a39-257b-40ad-beda-71796e2ac29e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.194307] env[62965]: DEBUG oslo_vmware.api [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 907.194307] env[62965]: value = "task-1389984" [ 907.194307] env[62965]: _type = "Task" [ 907.194307] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.201413] env[62965]: DEBUG oslo_vmware.api [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1389984, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.277599] env[62965]: DEBUG oslo_vmware.api [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389982, 'name': MoveVirtualDisk_Task} progress is 12%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.284664] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.660s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.285211] env[62965]: DEBUG nova.compute.manager [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 907.288060] env[62965]: DEBUG oslo_concurrency.lockutils [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.148s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.289570] env[62965]: INFO nova.compute.claims [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 907.421585] env[62965]: DEBUG oslo_vmware.api [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1389979, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.630151} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.421962] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] eed55d8e-21dc-47d4-a689-415ef6a3891d/eed55d8e-21dc-47d4-a689-415ef6a3891d.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 907.422241] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 907.422618] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-502b21e1-32a0-4bd5-b838-016c36a047d9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.432231] env[62965]: DEBUG oslo_vmware.api [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 907.432231] env[62965]: value = "task-1389985" [ 907.432231] env[62965]: _type = "Task" [ 907.432231] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.443695] env[62965]: DEBUG oslo_vmware.api [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1389985, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.455191] env[62965]: DEBUG nova.compute.manager [req-a6f965df-defd-4fed-8c55-e56b7984b22b req-8984d6a1-2f9a-4b7f-8d9d-f9472524f558 service nova] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Received event network-vif-plugged-fe1bf4cc-7773-462b-b8b5-1849c661d910 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 907.455475] env[62965]: DEBUG oslo_concurrency.lockutils [req-a6f965df-defd-4fed-8c55-e56b7984b22b req-8984d6a1-2f9a-4b7f-8d9d-f9472524f558 service nova] Acquiring lock "e0c4e188-a041-4aa7-95de-6e8718b17ade-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.455721] env[62965]: DEBUG oslo_concurrency.lockutils [req-a6f965df-defd-4fed-8c55-e56b7984b22b req-8984d6a1-2f9a-4b7f-8d9d-f9472524f558 service nova] Lock "e0c4e188-a041-4aa7-95de-6e8718b17ade-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.455988] env[62965]: DEBUG oslo_concurrency.lockutils [req-a6f965df-defd-4fed-8c55-e56b7984b22b req-8984d6a1-2f9a-4b7f-8d9d-f9472524f558 service nova] Lock "e0c4e188-a041-4aa7-95de-6e8718b17ade-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.456281] env[62965]: DEBUG nova.compute.manager [req-a6f965df-defd-4fed-8c55-e56b7984b22b req-8984d6a1-2f9a-4b7f-8d9d-f9472524f558 service nova] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] No waiting events found dispatching network-vif-plugged-fe1bf4cc-7773-462b-b8b5-1849c661d910 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 907.456725] env[62965]: WARNING nova.compute.manager [req-a6f965df-defd-4fed-8c55-e56b7984b22b req-8984d6a1-2f9a-4b7f-8d9d-f9472524f558 service nova] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Received unexpected event network-vif-plugged-fe1bf4cc-7773-462b-b8b5-1849c661d910 for instance with vm_state building and task_state spawning. [ 907.456725] env[62965]: DEBUG nova.compute.manager [req-a6f965df-defd-4fed-8c55-e56b7984b22b req-8984d6a1-2f9a-4b7f-8d9d-f9472524f558 service nova] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Received event network-changed-fe1bf4cc-7773-462b-b8b5-1849c661d910 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 907.456993] env[62965]: DEBUG nova.compute.manager [req-a6f965df-defd-4fed-8c55-e56b7984b22b req-8984d6a1-2f9a-4b7f-8d9d-f9472524f558 service nova] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Refreshing instance network info cache due to event network-changed-fe1bf4cc-7773-462b-b8b5-1849c661d910. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 907.457067] env[62965]: DEBUG oslo_concurrency.lockutils [req-a6f965df-defd-4fed-8c55-e56b7984b22b req-8984d6a1-2f9a-4b7f-8d9d-f9472524f558 service nova] Acquiring lock "refresh_cache-e0c4e188-a041-4aa7-95de-6e8718b17ade" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.533034] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a2f4cdb5-6a0d-4497-b977-13c7cb17c744 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "c2448b76-1553-4b68-a731-0bb0967f4c1d" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.533331] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a2f4cdb5-6a0d-4497-b977-13c7cb17c744 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "c2448b76-1553-4b68-a731-0bb0967f4c1d" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.533514] env[62965]: DEBUG nova.compute.manager [None req-a2f4cdb5-6a0d-4497-b977-13c7cb17c744 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 907.534451] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-159f7776-26ae-4151-8346-06bc8cb472eb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.542507] env[62965]: DEBUG nova.compute.manager [None req-a2f4cdb5-6a0d-4497-b977-13c7cb17c744 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62965) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 907.543166] env[62965]: DEBUG nova.objects.instance [None req-a2f4cdb5-6a0d-4497-b977-13c7cb17c744 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lazy-loading 'flavor' on Instance uuid c2448b76-1553-4b68-a731-0bb0967f4c1d {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 907.586076] env[62965]: DEBUG nova.network.neutron [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 907.709445] env[62965]: DEBUG oslo_vmware.api [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1389984, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.737785] env[62965]: DEBUG nova.network.neutron [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Updating instance_info_cache with network_info: [{"id": "fe1bf4cc-7773-462b-b8b5-1849c661d910", "address": "fa:16:3e:b8:4f:cc", "network": {"id": "a27e0199-89db-4b64-b3fd-4a12cd598e60", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-12171013-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d3edf0643c9d40e9bb568742b7a7a508", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f4a795c-8718-4a7c-aafe-9da231df10f8", "external-id": "nsx-vlan-transportzone-162", "segmentation_id": 162, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe1bf4cc-77", "ovs_interfaceid": "fe1bf4cc-7773-462b-b8b5-1849c661d910", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.778330] env[62965]: DEBUG oslo_vmware.api [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389982, 'name': MoveVirtualDisk_Task} progress is 32%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.794308] env[62965]: DEBUG nova.compute.utils [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 907.798936] env[62965]: DEBUG nova.compute.manager [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 907.799147] env[62965]: DEBUG nova.network.neutron [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 907.841032] env[62965]: DEBUG nova.policy [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '34e15ad3c87445e5a217300a23223540', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8bfa5cfdc7d5496482d87675f2a122dd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 907.947425] env[62965]: DEBUG oslo_vmware.api [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1389985, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073227} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.947752] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 907.948625] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dd22b0b-b783-4bc7-a857-daa946b7a4bf {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.974748] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] eed55d8e-21dc-47d4-a689-415ef6a3891d/eed55d8e-21dc-47d4-a689-415ef6a3891d.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 907.975445] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f899b3cb-8099-4306-9754-ce5bd656e2a0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.997710] env[62965]: DEBUG oslo_vmware.api [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 907.997710] env[62965]: value = "task-1389986" [ 907.997710] env[62965]: _type = "Task" [ 907.997710] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.009984] env[62965]: DEBUG oslo_vmware.api [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1389986, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.121331] env[62965]: DEBUG nova.network.neutron [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Successfully created port: 3edb59ed-a2b8-4d31-b975-9c9d64c3a22b {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 908.208460] env[62965]: DEBUG oslo_vmware.api [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1389984, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.240483] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Releasing lock "refresh_cache-e0c4e188-a041-4aa7-95de-6e8718b17ade" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.241374] env[62965]: DEBUG nova.compute.manager [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Instance network_info: |[{"id": "fe1bf4cc-7773-462b-b8b5-1849c661d910", "address": "fa:16:3e:b8:4f:cc", "network": {"id": "a27e0199-89db-4b64-b3fd-4a12cd598e60", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-12171013-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d3edf0643c9d40e9bb568742b7a7a508", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f4a795c-8718-4a7c-aafe-9da231df10f8", "external-id": "nsx-vlan-transportzone-162", "segmentation_id": 162, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe1bf4cc-77", "ovs_interfaceid": "fe1bf4cc-7773-462b-b8b5-1849c661d910", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 908.241823] env[62965]: DEBUG oslo_concurrency.lockutils [req-a6f965df-defd-4fed-8c55-e56b7984b22b req-8984d6a1-2f9a-4b7f-8d9d-f9472524f558 service nova] Acquired lock "refresh_cache-e0c4e188-a041-4aa7-95de-6e8718b17ade" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.242130] env[62965]: DEBUG nova.network.neutron [req-a6f965df-defd-4fed-8c55-e56b7984b22b req-8984d6a1-2f9a-4b7f-8d9d-f9472524f558 service nova] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Refreshing network info cache for port fe1bf4cc-7773-462b-b8b5-1849c661d910 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 908.243555] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b8:4f:cc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3f4a795c-8718-4a7c-aafe-9da231df10f8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fe1bf4cc-7773-462b-b8b5-1849c661d910', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 908.254536] env[62965]: DEBUG oslo.service.loopingcall [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 908.255794] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 908.256084] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1bc67055-9e41-45a6-800f-1eb5ce514d73 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.293258] env[62965]: DEBUG oslo_vmware.api [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389982, 'name': MoveVirtualDisk_Task} progress is 54%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.295487] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 908.295487] env[62965]: value = "task-1389987" [ 908.295487] env[62965]: _type = "Task" [ 908.295487] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.304033] env[62965]: DEBUG nova.compute.manager [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 908.318652] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389987, 'name': CreateVM_Task} progress is 10%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.515325] env[62965]: DEBUG oslo_vmware.api [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1389986, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.550448] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2f4cdb5-6a0d-4497-b977-13c7cb17c744 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 908.550883] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8511c30e-db43-4cbc-8287-2489626301ff {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.561978] env[62965]: DEBUG oslo_vmware.api [None req-a2f4cdb5-6a0d-4497-b977-13c7cb17c744 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 908.561978] env[62965]: value = "task-1389988" [ 908.561978] env[62965]: _type = "Task" [ 908.561978] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.574274] env[62965]: DEBUG oslo_vmware.api [None req-a2f4cdb5-6a0d-4497-b977-13c7cb17c744 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1389988, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.607945] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcce5156-9c2c-4c87-b174-20b88b7cb963 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.620693] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd74f34f-8c98-4192-8714-b1317dc8188e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.664376] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c024dc39-3399-4a37-bdb5-1dde2ea543d6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.679291] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efb90a0a-b746-4112-9bfd-654a23356a9a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.700635] env[62965]: DEBUG nova.compute.provider_tree [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 908.713323] env[62965]: DEBUG oslo_vmware.api [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1389984, 'name': PowerOnVM_Task, 'duration_secs': 1.261799} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.714194] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 908.714418] env[62965]: INFO nova.compute.manager [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Took 8.91 seconds to spawn the instance on the hypervisor. [ 908.714614] env[62965]: DEBUG nova.compute.manager [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 908.715514] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ae3128b-cbb0-4b48-9d32-3509a390b3ee {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.791682] env[62965]: DEBUG oslo_vmware.api [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389982, 'name': MoveVirtualDisk_Task} progress is 74%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.808585] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1389987, 'name': CreateVM_Task, 'duration_secs': 0.442635} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.808585] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 908.809139] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.809322] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.809658] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 908.813472] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba74871a-d506-4603-bcd6-6eb9d5684908 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.821458] env[62965]: DEBUG oslo_vmware.api [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 908.821458] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]528dee14-d9f1-30b1-60fe-77ee03115420" [ 908.821458] env[62965]: _type = "Task" [ 908.821458] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.832117] env[62965]: DEBUG oslo_vmware.api [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]528dee14-d9f1-30b1-60fe-77ee03115420, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.981263] env[62965]: DEBUG nova.network.neutron [req-a6f965df-defd-4fed-8c55-e56b7984b22b req-8984d6a1-2f9a-4b7f-8d9d-f9472524f558 service nova] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Updated VIF entry in instance network info cache for port fe1bf4cc-7773-462b-b8b5-1849c661d910. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 908.981644] env[62965]: DEBUG nova.network.neutron [req-a6f965df-defd-4fed-8c55-e56b7984b22b req-8984d6a1-2f9a-4b7f-8d9d-f9472524f558 service nova] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Updating instance_info_cache with network_info: [{"id": "fe1bf4cc-7773-462b-b8b5-1849c661d910", "address": "fa:16:3e:b8:4f:cc", "network": {"id": "a27e0199-89db-4b64-b3fd-4a12cd598e60", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-12171013-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d3edf0643c9d40e9bb568742b7a7a508", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f4a795c-8718-4a7c-aafe-9da231df10f8", "external-id": "nsx-vlan-transportzone-162", "segmentation_id": 162, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfe1bf4cc-77", "ovs_interfaceid": "fe1bf4cc-7773-462b-b8b5-1849c661d910", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.011246] env[62965]: DEBUG oslo_vmware.api [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1389986, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.076187] env[62965]: DEBUG oslo_vmware.api [None req-a2f4cdb5-6a0d-4497-b977-13c7cb17c744 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1389988, 'name': PowerOffVM_Task, 'duration_secs': 0.21147} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.076465] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2f4cdb5-6a0d-4497-b977-13c7cb17c744 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 909.076696] env[62965]: DEBUG nova.compute.manager [None req-a2f4cdb5-6a0d-4497-b977-13c7cb17c744 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 909.077581] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b553095b-ab07-49a8-ad6c-e3d63a072bed {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.095769] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c9895e6-9ec6-4bc3-8a85-a4cdc926665f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Volume attach. Driver type: vmdk {{(pid=62965) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 909.096031] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c9895e6-9ec6-4bc3-8a85-a4cdc926665f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-295006', 'volume_id': 'eeae4205-49ab-42e1-9d79-4b17b3886993', 'name': 'volume-eeae4205-49ab-42e1-9d79-4b17b3886993', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '64cf898c-2c97-4daa-80c8-0a1c252cbcf2', 'attached_at': '', 'detached_at': '', 'volume_id': 'eeae4205-49ab-42e1-9d79-4b17b3886993', 'serial': 'eeae4205-49ab-42e1-9d79-4b17b3886993'} {{(pid=62965) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 909.096974] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-410138ba-85f3-4c2a-aedd-d23ebc986f3d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.114919] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-740e7b09-0f65-43a1-9623-ca86270d28c9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.143987] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c9895e6-9ec6-4bc3-8a85-a4cdc926665f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] volume-eeae4205-49ab-42e1-9d79-4b17b3886993/volume-eeae4205-49ab-42e1-9d79-4b17b3886993.vmdk or device None with type thin {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 909.144361] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3d782333-920e-4a47-9f74-73a8ede582b1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.166447] env[62965]: DEBUG oslo_vmware.api [None req-7c9895e6-9ec6-4bc3-8a85-a4cdc926665f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Waiting for the task: (returnval){ [ 909.166447] env[62965]: value = "task-1389989" [ 909.166447] env[62965]: _type = "Task" [ 909.166447] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.178479] env[62965]: DEBUG oslo_vmware.api [None req-7c9895e6-9ec6-4bc3-8a85-a4cdc926665f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1389989, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.207731] env[62965]: DEBUG nova.scheduler.client.report [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 909.237978] env[62965]: INFO nova.compute.manager [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Took 18.57 seconds to build instance. [ 909.291724] env[62965]: DEBUG oslo_vmware.api [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389982, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.317581] env[62965]: DEBUG nova.compute.manager [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 909.336147] env[62965]: DEBUG oslo_vmware.api [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]528dee14-d9f1-30b1-60fe-77ee03115420, 'name': SearchDatastore_Task, 'duration_secs': 0.013629} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.338757] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.339112] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 909.339360] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.339523] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.339711] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 909.340610] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ce46143b-764d-43eb-b8d6-feb0296f1e49 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.346521] env[62965]: DEBUG nova.virt.hardware [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 909.346812] env[62965]: DEBUG nova.virt.hardware [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 909.347011] env[62965]: DEBUG nova.virt.hardware [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 909.347231] env[62965]: DEBUG nova.virt.hardware [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 909.347786] env[62965]: DEBUG nova.virt.hardware [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 909.347786] env[62965]: DEBUG nova.virt.hardware [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 909.347908] env[62965]: DEBUG nova.virt.hardware [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 909.348050] env[62965]: DEBUG nova.virt.hardware [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 909.348239] env[62965]: DEBUG nova.virt.hardware [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 909.348433] env[62965]: DEBUG nova.virt.hardware [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 909.348656] env[62965]: DEBUG nova.virt.hardware [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 909.349606] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4615a252-84bc-4f3c-ba86-4bf5b41d05d2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.353876] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 909.354111] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 909.355224] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c4f2e22-bf36-46de-b0ab-4faf438ccd02 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.361941] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97324998-9b9c-4219-803c-f40136e743bd {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.367737] env[62965]: DEBUG oslo_vmware.api [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 909.367737] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52802fcd-8deb-1d2f-3b3d-2fa50e007b20" [ 909.367737] env[62965]: _type = "Task" [ 909.367737] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.384871] env[62965]: DEBUG oslo_vmware.api [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52802fcd-8deb-1d2f-3b3d-2fa50e007b20, 'name': SearchDatastore_Task, 'duration_secs': 0.011499} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.385883] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60d342d6-0c51-419a-86ba-f6155ecff6f4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.391640] env[62965]: DEBUG oslo_vmware.api [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 909.391640] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52fb814d-720b-9602-de49-02b0c1b6723a" [ 909.391640] env[62965]: _type = "Task" [ 909.391640] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.400311] env[62965]: DEBUG oslo_vmware.api [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52fb814d-720b-9602-de49-02b0c1b6723a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.484202] env[62965]: DEBUG oslo_concurrency.lockutils [req-a6f965df-defd-4fed-8c55-e56b7984b22b req-8984d6a1-2f9a-4b7f-8d9d-f9472524f558 service nova] Releasing lock "refresh_cache-e0c4e188-a041-4aa7-95de-6e8718b17ade" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.509924] env[62965]: DEBUG oslo_vmware.api [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1389986, 'name': ReconfigVM_Task, 'duration_secs': 1.476842} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.509924] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Reconfigured VM instance instance-00000050 to attach disk [datastore1] eed55d8e-21dc-47d4-a689-415ef6a3891d/eed55d8e-21dc-47d4-a689-415ef6a3891d.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 909.510143] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-66450d5e-bc5c-45f7-9c04-086bb3594509 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.515788] env[62965]: DEBUG oslo_vmware.api [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 909.515788] env[62965]: value = "task-1389990" [ 909.515788] env[62965]: _type = "Task" [ 909.515788] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.529215] env[62965]: DEBUG oslo_vmware.api [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1389990, 'name': Rename_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.591653] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a2f4cdb5-6a0d-4497-b977-13c7cb17c744 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "c2448b76-1553-4b68-a731-0bb0967f4c1d" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.058s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.677105] env[62965]: DEBUG oslo_vmware.api [None req-7c9895e6-9ec6-4bc3-8a85-a4cdc926665f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1389989, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.712544] env[62965]: DEBUG oslo_concurrency.lockutils [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.424s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.713083] env[62965]: DEBUG nova.compute.manager [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 909.720017] env[62965]: DEBUG oslo_concurrency.lockutils [None req-322bb93e-e867-4da5-a1a8-6a5eae03a2a1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.043s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.720017] env[62965]: DEBUG nova.objects.instance [None req-322bb93e-e867-4da5-a1a8-6a5eae03a2a1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lazy-loading 'resources' on Instance uuid 8c478341-9b50-45de-9ba3-321e802ed1f0 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 909.740638] env[62965]: DEBUG oslo_concurrency.lockutils [None req-80f7090f-8903-45bc-880b-655192a3edee tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "892b1302-f536-4dcd-9227-a545ae1ac731" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.083s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.790875] env[62965]: DEBUG oslo_vmware.api [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389982, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.68608} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.791208] env[62965]: INFO nova.virt.vmwareapi.ds_util [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_9310edad-f829-4b40-8be5-438377eac181/OSTACK_IMG_9310edad-f829-4b40-8be5-438377eac181.vmdk to [datastore1] devstack-image-cache_base/7e436a80-d1ff-4c42-847a-9538961b1a6a/7e436a80-d1ff-4c42-847a-9538961b1a6a.vmdk. [ 909.791407] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Cleaning up location [datastore1] OSTACK_IMG_9310edad-f829-4b40-8be5-438377eac181 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 909.791574] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_9310edad-f829-4b40-8be5-438377eac181 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 909.792037] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7c272e9d-6331-4e8c-9143-d0157afa4b4f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.799472] env[62965]: DEBUG oslo_vmware.api [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 909.799472] env[62965]: value = "task-1389991" [ 909.799472] env[62965]: _type = "Task" [ 909.799472] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.809909] env[62965]: DEBUG oslo_vmware.api [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389991, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.874445] env[62965]: DEBUG nova.compute.manager [req-9a2e4b98-866f-487b-a3cc-d4137b202c1a req-d7d1266d-646d-41c3-b07c-ceeddff12c30 service nova] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Received event network-vif-plugged-3edb59ed-a2b8-4d31-b975-9c9d64c3a22b {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 909.874676] env[62965]: DEBUG oslo_concurrency.lockutils [req-9a2e4b98-866f-487b-a3cc-d4137b202c1a req-d7d1266d-646d-41c3-b07c-ceeddff12c30 service nova] Acquiring lock "fbbcbe60-92f3-497d-b190-520a2547b3a9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.874887] env[62965]: DEBUG oslo_concurrency.lockutils [req-9a2e4b98-866f-487b-a3cc-d4137b202c1a req-d7d1266d-646d-41c3-b07c-ceeddff12c30 service nova] Lock "fbbcbe60-92f3-497d-b190-520a2547b3a9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.875071] env[62965]: DEBUG oslo_concurrency.lockutils [req-9a2e4b98-866f-487b-a3cc-d4137b202c1a req-d7d1266d-646d-41c3-b07c-ceeddff12c30 service nova] Lock "fbbcbe60-92f3-497d-b190-520a2547b3a9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.875246] env[62965]: DEBUG nova.compute.manager [req-9a2e4b98-866f-487b-a3cc-d4137b202c1a req-d7d1266d-646d-41c3-b07c-ceeddff12c30 service nova] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] No waiting events found dispatching network-vif-plugged-3edb59ed-a2b8-4d31-b975-9c9d64c3a22b {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 909.875412] env[62965]: WARNING nova.compute.manager [req-9a2e4b98-866f-487b-a3cc-d4137b202c1a req-d7d1266d-646d-41c3-b07c-ceeddff12c30 service nova] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Received unexpected event network-vif-plugged-3edb59ed-a2b8-4d31-b975-9c9d64c3a22b for instance with vm_state building and task_state spawning. [ 909.901687] env[62965]: DEBUG oslo_vmware.api [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52fb814d-720b-9602-de49-02b0c1b6723a, 'name': SearchDatastore_Task, 'duration_secs': 0.015988} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.901946] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.902232] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] e0c4e188-a041-4aa7-95de-6e8718b17ade/e0c4e188-a041-4aa7-95de-6e8718b17ade.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 909.902490] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ffa7ecb7-99d2-4153-8491-cb9cd3d3dd9c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.908872] env[62965]: DEBUG oslo_vmware.api [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 909.908872] env[62965]: value = "task-1389992" [ 909.908872] env[62965]: _type = "Task" [ 909.908872] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.916619] env[62965]: DEBUG oslo_vmware.api [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1389992, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.025491] env[62965]: DEBUG oslo_concurrency.lockutils [None req-48f1ea8b-ebf5-4115-974e-b6c8f24e23b0 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "892b1302-f536-4dcd-9227-a545ae1ac731" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.025795] env[62965]: DEBUG oslo_concurrency.lockutils [None req-48f1ea8b-ebf5-4115-974e-b6c8f24e23b0 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "892b1302-f536-4dcd-9227-a545ae1ac731" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.026116] env[62965]: DEBUG oslo_concurrency.lockutils [None req-48f1ea8b-ebf5-4115-974e-b6c8f24e23b0 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "892b1302-f536-4dcd-9227-a545ae1ac731-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.026390] env[62965]: DEBUG oslo_concurrency.lockutils [None req-48f1ea8b-ebf5-4115-974e-b6c8f24e23b0 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "892b1302-f536-4dcd-9227-a545ae1ac731-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.026601] env[62965]: DEBUG oslo_concurrency.lockutils [None req-48f1ea8b-ebf5-4115-974e-b6c8f24e23b0 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "892b1302-f536-4dcd-9227-a545ae1ac731-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.028250] env[62965]: DEBUG oslo_vmware.api [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1389990, 'name': Rename_Task, 'duration_secs': 0.461795} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.028901] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 910.029400] env[62965]: INFO nova.compute.manager [None req-48f1ea8b-ebf5-4115-974e-b6c8f24e23b0 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Terminating instance [ 910.031639] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a59e4b43-2140-40e0-87d4-b51c01149646 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.038345] env[62965]: DEBUG oslo_vmware.api [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 910.038345] env[62965]: value = "task-1389993" [ 910.038345] env[62965]: _type = "Task" [ 910.038345] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.047973] env[62965]: DEBUG oslo_vmware.api [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1389993, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.112862] env[62965]: DEBUG nova.objects.instance [None req-a82fc0cc-bd37-4d94-9f90-1aacb8d30090 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lazy-loading 'flavor' on Instance uuid c2448b76-1553-4b68-a731-0bb0967f4c1d {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 910.176815] env[62965]: DEBUG oslo_vmware.api [None req-7c9895e6-9ec6-4bc3-8a85-a4cdc926665f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1389989, 'name': ReconfigVM_Task, 'duration_secs': 0.524447} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.177992] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c9895e6-9ec6-4bc3-8a85-a4cdc926665f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Reconfigured VM instance instance-00000040 to attach disk [datastore2] volume-eeae4205-49ab-42e1-9d79-4b17b3886993/volume-eeae4205-49ab-42e1-9d79-4b17b3886993.vmdk or device None with type thin {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 910.182161] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-02dc8e26-ecef-4f63-ac7f-5167f9dc9af2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.197797] env[62965]: DEBUG oslo_vmware.api [None req-7c9895e6-9ec6-4bc3-8a85-a4cdc926665f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Waiting for the task: (returnval){ [ 910.197797] env[62965]: value = "task-1389994" [ 910.197797] env[62965]: _type = "Task" [ 910.197797] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.206793] env[62965]: DEBUG oslo_vmware.api [None req-7c9895e6-9ec6-4bc3-8a85-a4cdc926665f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1389994, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.219162] env[62965]: DEBUG nova.compute.utils [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 910.223833] env[62965]: DEBUG nova.compute.manager [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 910.224014] env[62965]: DEBUG nova.network.neutron [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 910.263353] env[62965]: DEBUG nova.policy [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '182d38ca18c64bed8f3ecc3f95229756', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1952dedf340f4b7ba0d57af6bba9a749', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 910.313635] env[62965]: DEBUG oslo_vmware.api [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389991, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.036795} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.314016] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 910.314073] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Releasing lock "[datastore1] devstack-image-cache_base/7e436a80-d1ff-4c42-847a-9538961b1a6a/7e436a80-d1ff-4c42-847a-9538961b1a6a.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.314326] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/7e436a80-d1ff-4c42-847a-9538961b1a6a/7e436a80-d1ff-4c42-847a-9538961b1a6a.vmdk to [datastore1] f82d86ec-3364-449d-be54-35627082c02b/f82d86ec-3364-449d-be54-35627082c02b.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 910.314583] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9ffe38fc-9b72-416f-86e3-518afddc7074 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.321896] env[62965]: DEBUG nova.network.neutron [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Successfully updated port: 3edb59ed-a2b8-4d31-b975-9c9d64c3a22b {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 910.325392] env[62965]: DEBUG oslo_vmware.api [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 910.325392] env[62965]: value = "task-1389995" [ 910.325392] env[62965]: _type = "Task" [ 910.325392] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.338154] env[62965]: DEBUG oslo_vmware.api [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389995, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.348070] env[62965]: DEBUG nova.compute.manager [req-5bf7b368-ebbc-4cb3-8fd2-f61189343f47 req-d340826a-a1ab-424f-a27b-b147bed42a42 service nova] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Received event network-changed-3edb59ed-a2b8-4d31-b975-9c9d64c3a22b {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 910.348414] env[62965]: DEBUG nova.compute.manager [req-5bf7b368-ebbc-4cb3-8fd2-f61189343f47 req-d340826a-a1ab-424f-a27b-b147bed42a42 service nova] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Refreshing instance network info cache due to event network-changed-3edb59ed-a2b8-4d31-b975-9c9d64c3a22b. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 910.348582] env[62965]: DEBUG oslo_concurrency.lockutils [req-5bf7b368-ebbc-4cb3-8fd2-f61189343f47 req-d340826a-a1ab-424f-a27b-b147bed42a42 service nova] Acquiring lock "refresh_cache-fbbcbe60-92f3-497d-b190-520a2547b3a9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.348735] env[62965]: DEBUG oslo_concurrency.lockutils [req-5bf7b368-ebbc-4cb3-8fd2-f61189343f47 req-d340826a-a1ab-424f-a27b-b147bed42a42 service nova] Acquired lock "refresh_cache-fbbcbe60-92f3-497d-b190-520a2547b3a9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.348908] env[62965]: DEBUG nova.network.neutron [req-5bf7b368-ebbc-4cb3-8fd2-f61189343f47 req-d340826a-a1ab-424f-a27b-b147bed42a42 service nova] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Refreshing network info cache for port 3edb59ed-a2b8-4d31-b975-9c9d64c3a22b {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 910.422737] env[62965]: DEBUG oslo_vmware.api [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1389992, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.495911] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce7c0b9a-0cd5-45db-b665-b87b2ff7381f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.505089] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0397dbfa-db5f-48a3-9810-ef7ac49d0bdb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.539183] env[62965]: DEBUG nova.compute.manager [None req-48f1ea8b-ebf5-4115-974e-b6c8f24e23b0 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 910.539491] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-48f1ea8b-ebf5-4115-974e-b6c8f24e23b0 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 910.540686] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09432046-9087-4789-b8d3-5a2907f26723 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.544080] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a301eba9-57b3-44ec-bcce-673117779114 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.560819] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30fcb74f-8124-479f-bb4a-7cc2d4a94231 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.566031] env[62965]: DEBUG oslo_vmware.api [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1389993, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.566031] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-48f1ea8b-ebf5-4115-974e-b6c8f24e23b0 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 910.566031] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8fa05716-6110-4f65-8a8f-fad4f8c8743e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.580049] env[62965]: DEBUG nova.compute.provider_tree [None req-322bb93e-e867-4da5-a1a8-6a5eae03a2a1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 910.584139] env[62965]: DEBUG oslo_vmware.api [None req-48f1ea8b-ebf5-4115-974e-b6c8f24e23b0 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 910.584139] env[62965]: value = "task-1389996" [ 910.584139] env[62965]: _type = "Task" [ 910.584139] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.594719] env[62965]: DEBUG oslo_vmware.api [None req-48f1ea8b-ebf5-4115-974e-b6c8f24e23b0 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1389996, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.611626] env[62965]: DEBUG nova.network.neutron [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Successfully created port: 8fea9d40-d98f-4328-ad3e-ae708a849908 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 910.619125] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a82fc0cc-bd37-4d94-9f90-1aacb8d30090 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "refresh_cache-c2448b76-1553-4b68-a731-0bb0967f4c1d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.619285] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a82fc0cc-bd37-4d94-9f90-1aacb8d30090 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquired lock "refresh_cache-c2448b76-1553-4b68-a731-0bb0967f4c1d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.619519] env[62965]: DEBUG nova.network.neutron [None req-a82fc0cc-bd37-4d94-9f90-1aacb8d30090 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 910.619675] env[62965]: DEBUG nova.objects.instance [None req-a82fc0cc-bd37-4d94-9f90-1aacb8d30090 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lazy-loading 'info_cache' on Instance uuid c2448b76-1553-4b68-a731-0bb0967f4c1d {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 910.711108] env[62965]: DEBUG oslo_vmware.api [None req-7c9895e6-9ec6-4bc3-8a85-a4cdc926665f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1389994, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.724587] env[62965]: DEBUG nova.compute.manager [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 910.826322] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "refresh_cache-fbbcbe60-92f3-497d-b190-520a2547b3a9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.839479] env[62965]: DEBUG oslo_vmware.api [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389995, 'name': CopyVirtualDisk_Task} progress is 15%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.898178] env[62965]: DEBUG nova.network.neutron [req-5bf7b368-ebbc-4cb3-8fd2-f61189343f47 req-d340826a-a1ab-424f-a27b-b147bed42a42 service nova] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 910.921471] env[62965]: DEBUG oslo_vmware.api [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1389992, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.740119} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.921812] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] e0c4e188-a041-4aa7-95de-6e8718b17ade/e0c4e188-a041-4aa7-95de-6e8718b17ade.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 910.923027] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 910.923027] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-183adc6d-8b3f-4bb8-9fc4-386c02b9c3b2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.930853] env[62965]: DEBUG oslo_vmware.api [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 910.930853] env[62965]: value = "task-1389997" [ 910.930853] env[62965]: _type = "Task" [ 910.930853] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.941947] env[62965]: DEBUG oslo_vmware.api [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1389997, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.008595] env[62965]: DEBUG nova.network.neutron [req-5bf7b368-ebbc-4cb3-8fd2-f61189343f47 req-d340826a-a1ab-424f-a27b-b147bed42a42 service nova] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.052198] env[62965]: DEBUG oslo_vmware.api [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1389993, 'name': PowerOnVM_Task, 'duration_secs': 0.76187} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.052587] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 911.052874] env[62965]: INFO nova.compute.manager [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Took 8.87 seconds to spawn the instance on the hypervisor. [ 911.053086] env[62965]: DEBUG nova.compute.manager [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 911.054093] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0373f571-2ff7-465a-a41d-5cda352e58c6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.086898] env[62965]: DEBUG nova.scheduler.client.report [None req-322bb93e-e867-4da5-a1a8-6a5eae03a2a1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 911.102137] env[62965]: DEBUG oslo_vmware.api [None req-48f1ea8b-ebf5-4115-974e-b6c8f24e23b0 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1389996, 'name': PowerOffVM_Task, 'duration_secs': 0.342512} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.103192] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-48f1ea8b-ebf5-4115-974e-b6c8f24e23b0 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 911.103192] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-48f1ea8b-ebf5-4115-974e-b6c8f24e23b0 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 911.103192] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-30768677-6df3-408e-be88-83adfb2bc218 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.123954] env[62965]: DEBUG nova.objects.base [None req-a82fc0cc-bd37-4d94-9f90-1aacb8d30090 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=62965) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 911.184244] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-48f1ea8b-ebf5-4115-974e-b6c8f24e23b0 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 911.184244] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-48f1ea8b-ebf5-4115-974e-b6c8f24e23b0 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Deleting contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 911.184244] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-48f1ea8b-ebf5-4115-974e-b6c8f24e23b0 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Deleting the datastore file [datastore1] 892b1302-f536-4dcd-9227-a545ae1ac731 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 911.184244] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5f3feb65-6d34-4012-bbce-758d303090af {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.193599] env[62965]: DEBUG oslo_vmware.api [None req-48f1ea8b-ebf5-4115-974e-b6c8f24e23b0 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 911.193599] env[62965]: value = "task-1389999" [ 911.193599] env[62965]: _type = "Task" [ 911.193599] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.211552] env[62965]: DEBUG oslo_vmware.api [None req-7c9895e6-9ec6-4bc3-8a85-a4cdc926665f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1389994, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.215120] env[62965]: DEBUG oslo_vmware.api [None req-48f1ea8b-ebf5-4115-974e-b6c8f24e23b0 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1389999, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.337687] env[62965]: DEBUG oslo_vmware.api [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389995, 'name': CopyVirtualDisk_Task} progress is 35%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.443144] env[62965]: DEBUG oslo_vmware.api [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1389997, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.104726} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.443436] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 911.444261] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41098385-45cc-4b54-88a7-79f9fe15180e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.469118] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] e0c4e188-a041-4aa7-95de-6e8718b17ade/e0c4e188-a041-4aa7-95de-6e8718b17ade.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 911.469464] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ba0b9e36-6577-4927-82d2-77be7a57db10 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.489199] env[62965]: DEBUG oslo_vmware.api [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 911.489199] env[62965]: value = "task-1390000" [ 911.489199] env[62965]: _type = "Task" [ 911.489199] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.499514] env[62965]: DEBUG oslo_vmware.api [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390000, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.509889] env[62965]: DEBUG oslo_concurrency.lockutils [req-5bf7b368-ebbc-4cb3-8fd2-f61189343f47 req-d340826a-a1ab-424f-a27b-b147bed42a42 service nova] Releasing lock "refresh_cache-fbbcbe60-92f3-497d-b190-520a2547b3a9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.510237] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquired lock "refresh_cache-fbbcbe60-92f3-497d-b190-520a2547b3a9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.510407] env[62965]: DEBUG nova.network.neutron [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 911.573482] env[62965]: INFO nova.compute.manager [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Took 20.52 seconds to build instance. [ 911.591924] env[62965]: DEBUG oslo_concurrency.lockutils [None req-322bb93e-e867-4da5-a1a8-6a5eae03a2a1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.876s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.594553] env[62965]: DEBUG oslo_concurrency.lockutils [None req-57dfb0c1-c576-4f51-b7a4-f919972a1f07 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.804s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.594807] env[62965]: DEBUG nova.objects.instance [None req-57dfb0c1-c576-4f51-b7a4-f919972a1f07 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lazy-loading 'resources' on Instance uuid 4fc543a1-34ce-40bb-9e80-72ab9098eb38 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 911.618686] env[62965]: INFO nova.scheduler.client.report [None req-322bb93e-e867-4da5-a1a8-6a5eae03a2a1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Deleted allocations for instance 8c478341-9b50-45de-9ba3-321e802ed1f0 [ 911.707537] env[62965]: DEBUG oslo_vmware.api [None req-48f1ea8b-ebf5-4115-974e-b6c8f24e23b0 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1389999, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.715744] env[62965]: DEBUG oslo_vmware.api [None req-7c9895e6-9ec6-4bc3-8a85-a4cdc926665f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1389994, 'name': ReconfigVM_Task, 'duration_secs': 1.221077} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.716176] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c9895e6-9ec6-4bc3-8a85-a4cdc926665f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-295006', 'volume_id': 'eeae4205-49ab-42e1-9d79-4b17b3886993', 'name': 'volume-eeae4205-49ab-42e1-9d79-4b17b3886993', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '64cf898c-2c97-4daa-80c8-0a1c252cbcf2', 'attached_at': '', 'detached_at': '', 'volume_id': 'eeae4205-49ab-42e1-9d79-4b17b3886993', 'serial': 'eeae4205-49ab-42e1-9d79-4b17b3886993'} {{(pid=62965) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 911.735882] env[62965]: DEBUG nova.compute.manager [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 911.772792] env[62965]: DEBUG nova.virt.hardware [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 911.773069] env[62965]: DEBUG nova.virt.hardware [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 911.773235] env[62965]: DEBUG nova.virt.hardware [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 911.773423] env[62965]: DEBUG nova.virt.hardware [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 911.773601] env[62965]: DEBUG nova.virt.hardware [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 911.773758] env[62965]: DEBUG nova.virt.hardware [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 911.773994] env[62965]: DEBUG nova.virt.hardware [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 911.774250] env[62965]: DEBUG nova.virt.hardware [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 911.774447] env[62965]: DEBUG nova.virt.hardware [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 911.774723] env[62965]: DEBUG nova.virt.hardware [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 911.774810] env[62965]: DEBUG nova.virt.hardware [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 911.775757] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf6e1618-1abd-4508-8a28-99c97bd2d2de {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.789173] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c44bb9a8-627c-4855-a8ec-788fec26cd8a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.839676] env[62965]: DEBUG oslo_vmware.api [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389995, 'name': CopyVirtualDisk_Task} progress is 57%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.868597] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d6471311-1397-4d5d-875a-0d14109857fb tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquiring lock "eed55d8e-21dc-47d4-a689-415ef6a3891d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.891974] env[62965]: DEBUG nova.network.neutron [None req-a82fc0cc-bd37-4d94-9f90-1aacb8d30090 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Updating instance_info_cache with network_info: [{"id": "57672e8b-1df1-463d-b9b3-8290667d9fd2", "address": "fa:16:3e:53:46:47", "network": {"id": "4649fed7-c2ca-43a9-a183-7c99423da726", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1978698173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fab418e2fe1420793517663574b43bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43ad01d2-c7dd-453c-a929-8ad76294d13c", "external-id": "nsx-vlan-transportzone-176", "segmentation_id": 176, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57672e8b-1d", "ovs_interfaceid": "57672e8b-1df1-463d-b9b3-8290667d9fd2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.002621] env[62965]: DEBUG oslo_vmware.api [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390000, 'name': ReconfigVM_Task, 'duration_secs': 0.324531} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.003108] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Reconfigured VM instance instance-00000051 to attach disk [datastore2] e0c4e188-a041-4aa7-95de-6e8718b17ade/e0c4e188-a041-4aa7-95de-6e8718b17ade.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 912.003993] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ebbe3bd7-88eb-4cd2-8f3d-bcf39afec9cc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.015268] env[62965]: DEBUG oslo_vmware.api [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 912.015268] env[62965]: value = "task-1390001" [ 912.015268] env[62965]: _type = "Task" [ 912.015268] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.027056] env[62965]: DEBUG oslo_vmware.api [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390001, 'name': Rename_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.061977] env[62965]: DEBUG nova.network.neutron [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 912.075883] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9490c356-95cd-465a-be89-e37016d26266 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lock "eed55d8e-21dc-47d4-a689-415ef6a3891d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.033s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.076418] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d6471311-1397-4d5d-875a-0d14109857fb tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lock "eed55d8e-21dc-47d4-a689-415ef6a3891d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.208s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.076774] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d6471311-1397-4d5d-875a-0d14109857fb tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquiring lock "eed55d8e-21dc-47d4-a689-415ef6a3891d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.077152] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d6471311-1397-4d5d-875a-0d14109857fb tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lock "eed55d8e-21dc-47d4-a689-415ef6a3891d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.077431] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d6471311-1397-4d5d-875a-0d14109857fb tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lock "eed55d8e-21dc-47d4-a689-415ef6a3891d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.086210] env[62965]: INFO nova.compute.manager [None req-d6471311-1397-4d5d-875a-0d14109857fb tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Terminating instance [ 912.129840] env[62965]: DEBUG oslo_concurrency.lockutils [None req-322bb93e-e867-4da5-a1a8-6a5eae03a2a1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "8c478341-9b50-45de-9ba3-321e802ed1f0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.096s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.211491] env[62965]: DEBUG oslo_vmware.api [None req-48f1ea8b-ebf5-4115-974e-b6c8f24e23b0 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1389999, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.298950] env[62965]: DEBUG nova.network.neutron [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Updating instance_info_cache with network_info: [{"id": "3edb59ed-a2b8-4d31-b975-9c9d64c3a22b", "address": "fa:16:3e:92:26:4b", "network": {"id": "a5246cc1-6001-4c20-9724-fd36a0437582", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-644702631-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bfa5cfdc7d5496482d87675f2a122dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3edb59ed-a2", "ovs_interfaceid": "3edb59ed-a2b8-4d31-b975-9c9d64c3a22b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.342264] env[62965]: DEBUG oslo_vmware.api [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389995, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.393084] env[62965]: DEBUG nova.network.neutron [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Successfully updated port: 8fea9d40-d98f-4328-ad3e-ae708a849908 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 912.394833] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a82fc0cc-bd37-4d94-9f90-1aacb8d30090 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Releasing lock "refresh_cache-c2448b76-1553-4b68-a731-0bb0967f4c1d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.396689] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e09191d-4cde-4805-9530-ebcc2cc526ee {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.411038] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f7aea6f-77ac-4306-8a91-6dbe579f14d3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.453329] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-145f6bc6-314d-499a-9b83-93ce15eefd14 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.464094] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79b04c5b-7661-48e0-8450-d9a0b0ce2183 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.480992] env[62965]: DEBUG nova.compute.provider_tree [None req-57dfb0c1-c576-4f51-b7a4-f919972a1f07 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 912.511394] env[62965]: DEBUG nova.compute.manager [req-51a0529c-51df-4264-9dfa-c54750d998d6 req-6daadd7e-06b8-4c27-8b88-5431fbe7c326 service nova] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Received event network-vif-plugged-8fea9d40-d98f-4328-ad3e-ae708a849908 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 912.511599] env[62965]: DEBUG oslo_concurrency.lockutils [req-51a0529c-51df-4264-9dfa-c54750d998d6 req-6daadd7e-06b8-4c27-8b88-5431fbe7c326 service nova] Acquiring lock "00f24976-2f0e-4a2f-8f2e-012b0fce479f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.511808] env[62965]: DEBUG oslo_concurrency.lockutils [req-51a0529c-51df-4264-9dfa-c54750d998d6 req-6daadd7e-06b8-4c27-8b88-5431fbe7c326 service nova] Lock "00f24976-2f0e-4a2f-8f2e-012b0fce479f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.511973] env[62965]: DEBUG oslo_concurrency.lockutils [req-51a0529c-51df-4264-9dfa-c54750d998d6 req-6daadd7e-06b8-4c27-8b88-5431fbe7c326 service nova] Lock "00f24976-2f0e-4a2f-8f2e-012b0fce479f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.512160] env[62965]: DEBUG nova.compute.manager [req-51a0529c-51df-4264-9dfa-c54750d998d6 req-6daadd7e-06b8-4c27-8b88-5431fbe7c326 service nova] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] No waiting events found dispatching network-vif-plugged-8fea9d40-d98f-4328-ad3e-ae708a849908 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 912.512730] env[62965]: WARNING nova.compute.manager [req-51a0529c-51df-4264-9dfa-c54750d998d6 req-6daadd7e-06b8-4c27-8b88-5431fbe7c326 service nova] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Received unexpected event network-vif-plugged-8fea9d40-d98f-4328-ad3e-ae708a849908 for instance with vm_state building and task_state spawning. [ 912.521015] env[62965]: DEBUG oslo_concurrency.lockutils [None req-624ecc99-34aa-4f6d-b48a-9805c8d1ab8f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquiring lock "6def4a38-d3b8-4724-97a1-1859b239d900" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.521269] env[62965]: DEBUG oslo_concurrency.lockutils [None req-624ecc99-34aa-4f6d-b48a-9805c8d1ab8f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "6def4a38-d3b8-4724-97a1-1859b239d900" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.530763] env[62965]: DEBUG oslo_vmware.api [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390001, 'name': Rename_Task, 'duration_secs': 0.158873} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.531612] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 912.531875] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0a310fe5-f9a6-45de-8f6a-0b6c2c0b1e79 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.541428] env[62965]: DEBUG oslo_vmware.api [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 912.541428] env[62965]: value = "task-1390002" [ 912.541428] env[62965]: _type = "Task" [ 912.541428] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.552996] env[62965]: DEBUG oslo_vmware.api [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390002, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.592430] env[62965]: DEBUG nova.compute.manager [None req-d6471311-1397-4d5d-875a-0d14109857fb tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 912.593044] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-d6471311-1397-4d5d-875a-0d14109857fb tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 912.593632] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08c8f2e6-c416-416e-ad66-4a518305a86f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.604148] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6471311-1397-4d5d-875a-0d14109857fb tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 912.604484] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aa668b65-ad9f-4594-a121-6e606ec43863 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.613621] env[62965]: DEBUG oslo_vmware.api [None req-d6471311-1397-4d5d-875a-0d14109857fb tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 912.613621] env[62965]: value = "task-1390003" [ 912.613621] env[62965]: _type = "Task" [ 912.613621] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.623832] env[62965]: DEBUG oslo_vmware.api [None req-d6471311-1397-4d5d-875a-0d14109857fb tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390003, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.706833] env[62965]: DEBUG oslo_vmware.api [None req-48f1ea8b-ebf5-4115-974e-b6c8f24e23b0 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1389999, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.773179] env[62965]: DEBUG nova.objects.instance [None req-7c9895e6-9ec6-4bc3-8a85-a4cdc926665f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Lazy-loading 'flavor' on Instance uuid 64cf898c-2c97-4daa-80c8-0a1c252cbcf2 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 912.805314] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Releasing lock "refresh_cache-fbbcbe60-92f3-497d-b190-520a2547b3a9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.805646] env[62965]: DEBUG nova.compute.manager [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Instance network_info: |[{"id": "3edb59ed-a2b8-4d31-b975-9c9d64c3a22b", "address": "fa:16:3e:92:26:4b", "network": {"id": "a5246cc1-6001-4c20-9724-fd36a0437582", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-644702631-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bfa5cfdc7d5496482d87675f2a122dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3edb59ed-a2", "ovs_interfaceid": "3edb59ed-a2b8-4d31-b975-9c9d64c3a22b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 912.806395] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:92:26:4b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '61b8f0db-488e-42d7-bf6c-6c1665cd5616', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3edb59ed-a2b8-4d31-b975-9c9d64c3a22b', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 912.813975] env[62965]: DEBUG oslo.service.loopingcall [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 912.814601] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 912.814844] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cffe341b-e4d4-4e38-b738-f715b52d26ee {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.837391] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 912.837391] env[62965]: value = "task-1390004" [ 912.837391] env[62965]: _type = "Task" [ 912.837391] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.843881] env[62965]: DEBUG oslo_vmware.api [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389995, 'name': CopyVirtualDisk_Task} progress is 97%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.851105] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390004, 'name': CreateVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.902615] env[62965]: DEBUG oslo_concurrency.lockutils [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "refresh_cache-00f24976-2f0e-4a2f-8f2e-012b0fce479f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.902752] env[62965]: DEBUG oslo_concurrency.lockutils [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquired lock "refresh_cache-00f24976-2f0e-4a2f-8f2e-012b0fce479f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.902899] env[62965]: DEBUG nova.network.neutron [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 913.014352] env[62965]: DEBUG nova.scheduler.client.report [None req-57dfb0c1-c576-4f51-b7a4-f919972a1f07 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Updated inventory for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with generation 92 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 913.014633] env[62965]: DEBUG nova.compute.provider_tree [None req-57dfb0c1-c576-4f51-b7a4-f919972a1f07 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Updating resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 generation from 92 to 93 during operation: update_inventory {{(pid=62965) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 913.014818] env[62965]: DEBUG nova.compute.provider_tree [None req-57dfb0c1-c576-4f51-b7a4-f919972a1f07 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 913.024721] env[62965]: DEBUG nova.compute.utils [None req-624ecc99-34aa-4f6d-b48a-9805c8d1ab8f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 913.052669] env[62965]: DEBUG oslo_vmware.api [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390002, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.122965] env[62965]: DEBUG oslo_vmware.api [None req-d6471311-1397-4d5d-875a-0d14109857fb tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390003, 'name': PowerOffVM_Task, 'duration_secs': 0.415925} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.123285] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6471311-1397-4d5d-875a-0d14109857fb tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 913.123455] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-d6471311-1397-4d5d-875a-0d14109857fb tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 913.123711] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-64273958-0247-462e-b5e2-51d1498fdc22 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.205761] env[62965]: DEBUG oslo_vmware.api [None req-48f1ea8b-ebf5-4115-974e-b6c8f24e23b0 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1389999, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.762017} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.206037] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-48f1ea8b-ebf5-4115-974e-b6c8f24e23b0 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 913.206227] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-48f1ea8b-ebf5-4115-974e-b6c8f24e23b0 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Deleted contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 913.206766] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-48f1ea8b-ebf5-4115-974e-b6c8f24e23b0 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 913.206766] env[62965]: INFO nova.compute.manager [None req-48f1ea8b-ebf5-4115-974e-b6c8f24e23b0 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Took 2.67 seconds to destroy the instance on the hypervisor. [ 913.206884] env[62965]: DEBUG oslo.service.loopingcall [None req-48f1ea8b-ebf5-4115-974e-b6c8f24e23b0 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 913.207171] env[62965]: DEBUG nova.compute.manager [-] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 913.207268] env[62965]: DEBUG nova.network.neutron [-] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 913.210249] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-d6471311-1397-4d5d-875a-0d14109857fb tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 913.210371] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-d6471311-1397-4d5d-875a-0d14109857fb tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Deleting contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 913.210543] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6471311-1397-4d5d-875a-0d14109857fb tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Deleting the datastore file [datastore1] eed55d8e-21dc-47d4-a689-415ef6a3891d {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 913.211129] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6a9696aa-0420-4f6c-97fa-95bf8aa20684 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.217249] env[62965]: DEBUG oslo_vmware.api [None req-d6471311-1397-4d5d-875a-0d14109857fb tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 913.217249] env[62965]: value = "task-1390006" [ 913.217249] env[62965]: _type = "Task" [ 913.217249] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.225317] env[62965]: DEBUG oslo_vmware.api [None req-d6471311-1397-4d5d-875a-0d14109857fb tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390006, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.279671] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c9895e6-9ec6-4bc3-8a85-a4cdc926665f tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Lock "64cf898c-2c97-4daa-80c8-0a1c252cbcf2" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.804s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.340190] env[62965]: DEBUG oslo_vmware.api [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1389995, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.607156} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.343443] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/7e436a80-d1ff-4c42-847a-9538961b1a6a/7e436a80-d1ff-4c42-847a-9538961b1a6a.vmdk to [datastore1] f82d86ec-3364-449d-be54-35627082c02b/f82d86ec-3364-449d-be54-35627082c02b.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 913.344327] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57c499c5-02c7-466d-bae7-49a216b2e849 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.352316] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390004, 'name': CreateVM_Task} progress is 99%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.370064] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] f82d86ec-3364-449d-be54-35627082c02b/f82d86ec-3364-449d-be54-35627082c02b.vmdk or device None with type streamOptimized {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 913.370714] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-18ce4672-1abb-496e-8bec-b2847b99db87 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.389792] env[62965]: DEBUG oslo_vmware.api [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 913.389792] env[62965]: value = "task-1390007" [ 913.389792] env[62965]: _type = "Task" [ 913.389792] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.398339] env[62965]: DEBUG oslo_vmware.api [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390007, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.405080] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-a82fc0cc-bd37-4d94-9f90-1aacb8d30090 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 913.405394] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7c94dcbd-1988-46dd-ae3d-f7143c3913ea {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.412927] env[62965]: DEBUG oslo_vmware.api [None req-a82fc0cc-bd37-4d94-9f90-1aacb8d30090 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 913.412927] env[62965]: value = "task-1390008" [ 913.412927] env[62965]: _type = "Task" [ 913.412927] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.425037] env[62965]: DEBUG oslo_vmware.api [None req-a82fc0cc-bd37-4d94-9f90-1aacb8d30090 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390008, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.457517] env[62965]: DEBUG nova.network.neutron [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 913.520111] env[62965]: DEBUG oslo_concurrency.lockutils [None req-57dfb0c1-c576-4f51-b7a4-f919972a1f07 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.925s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.525066] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ba31b606-70ea-44b4-80d5-2d58061ebaa5 tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.764s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.525066] env[62965]: DEBUG nova.objects.instance [None req-ba31b606-70ea-44b4-80d5-2d58061ebaa5 tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Lazy-loading 'resources' on Instance uuid fa511b61-887e-46a1-aaf6-586d716c0fb5 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 913.526878] env[62965]: DEBUG oslo_concurrency.lockutils [None req-624ecc99-34aa-4f6d-b48a-9805c8d1ab8f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "6def4a38-d3b8-4724-97a1-1859b239d900" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.005s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.555561] env[62965]: INFO nova.scheduler.client.report [None req-57dfb0c1-c576-4f51-b7a4-f919972a1f07 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Deleted allocations for instance 4fc543a1-34ce-40bb-9e80-72ab9098eb38 [ 913.559836] env[62965]: DEBUG oslo_vmware.api [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390002, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.645404] env[62965]: DEBUG nova.network.neutron [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Updating instance_info_cache with network_info: [{"id": "8fea9d40-d98f-4328-ad3e-ae708a849908", "address": "fa:16:3e:c3:8c:35", "network": {"id": "a9699d8b-5e76-4666-a236-33ab8f22345e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1329540850-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1952dedf340f4b7ba0d57af6bba9a749", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a316376e-2ef0-4b1e-b40c-10321ebd7e1a", "external-id": "nsx-vlan-transportzone-942", "segmentation_id": 942, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8fea9d40-d9", "ovs_interfaceid": "8fea9d40-d98f-4328-ad3e-ae708a849908", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.728940] env[62965]: DEBUG oslo_vmware.api [None req-d6471311-1397-4d5d-875a-0d14109857fb tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390006, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.184716} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.729281] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6471311-1397-4d5d-875a-0d14109857fb tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 913.729515] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-d6471311-1397-4d5d-875a-0d14109857fb tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Deleted contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 913.729741] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-d6471311-1397-4d5d-875a-0d14109857fb tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 913.730019] env[62965]: INFO nova.compute.manager [None req-d6471311-1397-4d5d-875a-0d14109857fb tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Took 1.14 seconds to destroy the instance on the hypervisor. [ 913.730311] env[62965]: DEBUG oslo.service.loopingcall [None req-d6471311-1397-4d5d-875a-0d14109857fb tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 913.730548] env[62965]: DEBUG nova.compute.manager [-] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 913.730665] env[62965]: DEBUG nova.network.neutron [-] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 913.850319] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390004, 'name': CreateVM_Task, 'duration_secs': 0.703517} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.850434] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 913.851953] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.852152] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.852482] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 913.852739] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-352da4ec-4b80-4c23-8952-aaddec4c8e3e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.857839] env[62965]: DEBUG oslo_vmware.api [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 913.857839] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52cfcfb1-996a-7f6e-f5b7-cad9dfa0075f" [ 913.857839] env[62965]: _type = "Task" [ 913.857839] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.865670] env[62965]: DEBUG oslo_vmware.api [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52cfcfb1-996a-7f6e-f5b7-cad9dfa0075f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.899630] env[62965]: DEBUG oslo_vmware.api [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390007, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.923902] env[62965]: DEBUG oslo_vmware.api [None req-a82fc0cc-bd37-4d94-9f90-1aacb8d30090 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390008, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.925170] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a3082e98-a878-445b-a2b1-07e8cce5eb80 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Acquiring lock "64cf898c-2c97-4daa-80c8-0a1c252cbcf2" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.925407] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a3082e98-a878-445b-a2b1-07e8cce5eb80 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Lock "64cf898c-2c97-4daa-80c8-0a1c252cbcf2" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.053688] env[62965]: DEBUG oslo_vmware.api [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390002, 'name': PowerOnVM_Task, 'duration_secs': 1.091515} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.056119] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 914.056338] env[62965]: INFO nova.compute.manager [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Took 8.37 seconds to spawn the instance on the hypervisor. [ 914.056522] env[62965]: DEBUG nova.compute.manager [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 914.057460] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5761d159-8026-4249-ae99-eedba12f9dd5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.070516] env[62965]: DEBUG oslo_concurrency.lockutils [None req-57dfb0c1-c576-4f51-b7a4-f919972a1f07 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "4fc543a1-34ce-40bb-9e80-72ab9098eb38" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.784s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.081687] env[62965]: DEBUG nova.network.neutron [-] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.148043] env[62965]: DEBUG oslo_concurrency.lockutils [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Releasing lock "refresh_cache-00f24976-2f0e-4a2f-8f2e-012b0fce479f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.148043] env[62965]: DEBUG nova.compute.manager [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Instance network_info: |[{"id": "8fea9d40-d98f-4328-ad3e-ae708a849908", "address": "fa:16:3e:c3:8c:35", "network": {"id": "a9699d8b-5e76-4666-a236-33ab8f22345e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1329540850-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1952dedf340f4b7ba0d57af6bba9a749", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a316376e-2ef0-4b1e-b40c-10321ebd7e1a", "external-id": "nsx-vlan-transportzone-942", "segmentation_id": 942, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8fea9d40-d9", "ovs_interfaceid": "8fea9d40-d98f-4328-ad3e-ae708a849908", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 914.149013] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c3:8c:35', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a316376e-2ef0-4b1e-b40c-10321ebd7e1a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8fea9d40-d98f-4328-ad3e-ae708a849908', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 914.157027] env[62965]: DEBUG oslo.service.loopingcall [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 914.159630] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 914.160101] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7317cd55-9f0b-4b63-8c17-74d7bd6a0200 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.185832] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 914.185832] env[62965]: value = "task-1390009" [ 914.185832] env[62965]: _type = "Task" [ 914.185832] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.196993] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390009, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.252766] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-635cefbc-070c-4b73-a04e-bdd88192f89a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.260304] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b32f49be-e7dd-4ea3-b2f1-be421fce8ff0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.292247] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c1b9708-bb76-4d14-8b23-3d4b17611922 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.299733] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-697c2757-21b0-4086-bc29-da29bc1450f3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.313909] env[62965]: DEBUG nova.compute.provider_tree [None req-ba31b606-70ea-44b4-80d5-2d58061ebaa5 tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 914.368429] env[62965]: DEBUG oslo_vmware.api [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52cfcfb1-996a-7f6e-f5b7-cad9dfa0075f, 'name': SearchDatastore_Task, 'duration_secs': 0.009533} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.368721] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.368969] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 914.369243] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.369512] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.369599] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 914.369895] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3445c68e-0da7-492c-9418-8241be5b6db1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.378568] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 914.378759] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 914.379725] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0012ae09-e7e0-4ffc-aede-0b527545c737 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.384850] env[62965]: DEBUG oslo_vmware.api [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 914.384850] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]527c6789-1eb9-1b86-9073-f1e2fed0fdd2" [ 914.384850] env[62965]: _type = "Task" [ 914.384850] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.393367] env[62965]: DEBUG oslo_vmware.api [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]527c6789-1eb9-1b86-9073-f1e2fed0fdd2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.401325] env[62965]: DEBUG oslo_vmware.api [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390007, 'name': ReconfigVM_Task, 'duration_secs': 0.622494} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.401608] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Reconfigured VM instance instance-0000004e to attach disk [datastore1] f82d86ec-3364-449d-be54-35627082c02b/f82d86ec-3364-449d-be54-35627082c02b.vmdk or device None with type streamOptimized {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 914.402340] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2c4fc121-1a4a-4d13-9daf-7fccf48b9af6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.408816] env[62965]: DEBUG oslo_vmware.api [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 914.408816] env[62965]: value = "task-1390010" [ 914.408816] env[62965]: _type = "Task" [ 914.408816] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.416974] env[62965]: DEBUG oslo_vmware.api [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390010, 'name': Rename_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.424116] env[62965]: DEBUG oslo_vmware.api [None req-a82fc0cc-bd37-4d94-9f90-1aacb8d30090 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390008, 'name': PowerOnVM_Task, 'duration_secs': 0.613479} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.424361] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-a82fc0cc-bd37-4d94-9f90-1aacb8d30090 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 914.424578] env[62965]: DEBUG nova.compute.manager [None req-a82fc0cc-bd37-4d94-9f90-1aacb8d30090 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 914.425322] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d83f82c-2e16-4924-959f-365d6911e51a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.428430] env[62965]: DEBUG nova.compute.utils [None req-a3082e98-a878-445b-a2b1-07e8cce5eb80 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 914.459883] env[62965]: DEBUG nova.network.neutron [-] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.577506] env[62965]: INFO nova.compute.manager [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Took 18.39 seconds to build instance. [ 914.586753] env[62965]: INFO nova.compute.manager [-] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Took 1.38 seconds to deallocate network for instance. [ 914.587411] env[62965]: DEBUG oslo_concurrency.lockutils [None req-624ecc99-34aa-4f6d-b48a-9805c8d1ab8f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquiring lock "6def4a38-d3b8-4724-97a1-1859b239d900" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.587658] env[62965]: DEBUG oslo_concurrency.lockutils [None req-624ecc99-34aa-4f6d-b48a-9805c8d1ab8f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "6def4a38-d3b8-4724-97a1-1859b239d900" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.587885] env[62965]: INFO nova.compute.manager [None req-624ecc99-34aa-4f6d-b48a-9805c8d1ab8f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Attaching volume edb73cdb-c23b-4dae-a3b5-8f4c025464d7 to /dev/sdb [ 914.614715] env[62965]: DEBUG nova.compute.manager [req-e97cdcfb-a83d-4598-9037-56d444a142c9 req-cdd8865a-30d4-42e5-8e15-cf8faae44b8e service nova] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Received event network-changed-8fea9d40-d98f-4328-ad3e-ae708a849908 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 914.614915] env[62965]: DEBUG nova.compute.manager [req-e97cdcfb-a83d-4598-9037-56d444a142c9 req-cdd8865a-30d4-42e5-8e15-cf8faae44b8e service nova] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Refreshing instance network info cache due to event network-changed-8fea9d40-d98f-4328-ad3e-ae708a849908. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 914.615149] env[62965]: DEBUG oslo_concurrency.lockutils [req-e97cdcfb-a83d-4598-9037-56d444a142c9 req-cdd8865a-30d4-42e5-8e15-cf8faae44b8e service nova] Acquiring lock "refresh_cache-00f24976-2f0e-4a2f-8f2e-012b0fce479f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.615293] env[62965]: DEBUG oslo_concurrency.lockutils [req-e97cdcfb-a83d-4598-9037-56d444a142c9 req-cdd8865a-30d4-42e5-8e15-cf8faae44b8e service nova] Acquired lock "refresh_cache-00f24976-2f0e-4a2f-8f2e-012b0fce479f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.615456] env[62965]: DEBUG nova.network.neutron [req-e97cdcfb-a83d-4598-9037-56d444a142c9 req-cdd8865a-30d4-42e5-8e15-cf8faae44b8e service nova] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Refreshing network info cache for port 8fea9d40-d98f-4328-ad3e-ae708a849908 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 914.626156] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d0b449a-bf11-4906-b312-5ffda43ed8a4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.633823] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd3bee14-8ace-48ad-ac40-5540c43def9b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.647193] env[62965]: DEBUG nova.virt.block_device [None req-624ecc99-34aa-4f6d-b48a-9805c8d1ab8f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Updating existing volume attachment record: 6cedb2fd-8d9c-4128-b0b2-2396b0f993a8 {{(pid=62965) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 914.694859] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390009, 'name': CreateVM_Task, 'duration_secs': 0.316361} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.694946] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 914.695646] env[62965]: DEBUG oslo_concurrency.lockutils [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.695800] env[62965]: DEBUG oslo_concurrency.lockutils [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.696155] env[62965]: DEBUG oslo_concurrency.lockutils [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 914.696730] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9fbddd08-a785-462b-939d-f4ff215c6778 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.701042] env[62965]: DEBUG oslo_vmware.api [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Waiting for the task: (returnval){ [ 914.701042] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52168ff1-a1af-e00d-6034-77d20c8166e7" [ 914.701042] env[62965]: _type = "Task" [ 914.701042] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.709091] env[62965]: DEBUG oslo_vmware.api [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52168ff1-a1af-e00d-6034-77d20c8166e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.817216] env[62965]: DEBUG nova.scheduler.client.report [None req-ba31b606-70ea-44b4-80d5-2d58061ebaa5 tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 914.895014] env[62965]: DEBUG oslo_vmware.api [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]527c6789-1eb9-1b86-9073-f1e2fed0fdd2, 'name': SearchDatastore_Task, 'duration_secs': 0.010911} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.897685] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ebf53de-c901-4c28-9f3a-3d6ea83a5498 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.901219] env[62965]: DEBUG oslo_vmware.api [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 914.901219] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5293835e-7578-4f25-6ea8-00e35481ddea" [ 914.901219] env[62965]: _type = "Task" [ 914.901219] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.908901] env[62965]: DEBUG oslo_vmware.api [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5293835e-7578-4f25-6ea8-00e35481ddea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.916320] env[62965]: DEBUG oslo_vmware.api [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390010, 'name': Rename_Task, 'duration_secs': 0.174249} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.916576] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 914.916838] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-db57ea90-63cb-465d-9a2b-10d110e5c8f8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.921591] env[62965]: DEBUG oslo_vmware.api [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 914.921591] env[62965]: value = "task-1390014" [ 914.921591] env[62965]: _type = "Task" [ 914.921591] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.928662] env[62965]: DEBUG oslo_vmware.api [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390014, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.933341] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a3082e98-a878-445b-a2b1-07e8cce5eb80 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Lock "64cf898c-2c97-4daa-80c8-0a1c252cbcf2" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.962113] env[62965]: INFO nova.compute.manager [-] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Took 1.23 seconds to deallocate network for instance. [ 915.079099] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a073745d-8d6b-46db-b3e1-41f53c7a3c40 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Lock "e0c4e188-a041-4aa7-95de-6e8718b17ade" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.900s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.100252] env[62965]: DEBUG oslo_concurrency.lockutils [None req-48f1ea8b-ebf5-4115-974e-b6c8f24e23b0 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.213074] env[62965]: DEBUG oslo_vmware.api [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52168ff1-a1af-e00d-6034-77d20c8166e7, 'name': SearchDatastore_Task, 'duration_secs': 0.011305} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.213395] env[62965]: DEBUG oslo_concurrency.lockutils [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.213627] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 915.213843] env[62965]: DEBUG oslo_concurrency.lockutils [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.315089] env[62965]: DEBUG nova.network.neutron [req-e97cdcfb-a83d-4598-9037-56d444a142c9 req-cdd8865a-30d4-42e5-8e15-cf8faae44b8e service nova] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Updated VIF entry in instance network info cache for port 8fea9d40-d98f-4328-ad3e-ae708a849908. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 915.315659] env[62965]: DEBUG nova.network.neutron [req-e97cdcfb-a83d-4598-9037-56d444a142c9 req-cdd8865a-30d4-42e5-8e15-cf8faae44b8e service nova] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Updating instance_info_cache with network_info: [{"id": "8fea9d40-d98f-4328-ad3e-ae708a849908", "address": "fa:16:3e:c3:8c:35", "network": {"id": "a9699d8b-5e76-4666-a236-33ab8f22345e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1329540850-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1952dedf340f4b7ba0d57af6bba9a749", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a316376e-2ef0-4b1e-b40c-10321ebd7e1a", "external-id": "nsx-vlan-transportzone-942", "segmentation_id": 942, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8fea9d40-d9", "ovs_interfaceid": "8fea9d40-d98f-4328-ad3e-ae708a849908", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.322550] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ba31b606-70ea-44b4-80d5-2d58061ebaa5 tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.800s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.326120] env[62965]: DEBUG oslo_concurrency.lockutils [None req-48f1ea8b-ebf5-4115-974e-b6c8f24e23b0 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.226s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.326401] env[62965]: DEBUG nova.objects.instance [None req-48f1ea8b-ebf5-4115-974e-b6c8f24e23b0 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lazy-loading 'resources' on Instance uuid 892b1302-f536-4dcd-9227-a545ae1ac731 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 915.342062] env[62965]: INFO nova.scheduler.client.report [None req-ba31b606-70ea-44b4-80d5-2d58061ebaa5 tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Deleted allocations for instance fa511b61-887e-46a1-aaf6-586d716c0fb5 [ 915.412276] env[62965]: DEBUG oslo_vmware.api [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5293835e-7578-4f25-6ea8-00e35481ddea, 'name': SearchDatastore_Task, 'duration_secs': 0.010936} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.412276] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.412489] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] fbbcbe60-92f3-497d-b190-520a2547b3a9/fbbcbe60-92f3-497d-b190-520a2547b3a9.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 915.412857] env[62965]: DEBUG oslo_concurrency.lockutils [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.412942] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 915.413174] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-492081d8-df0c-492c-afa4-95e10fe6c87d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.415270] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fe3ef257-894b-40bb-b4da-8571e9aeff83 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.421916] env[62965]: DEBUG oslo_vmware.api [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 915.421916] env[62965]: value = "task-1390015" [ 915.421916] env[62965]: _type = "Task" [ 915.421916] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.429205] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 915.429504] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 915.430861] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-63e72e0f-5e41-44a3-945b-285da68bc307 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.438812] env[62965]: DEBUG oslo_vmware.api [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390015, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.439083] env[62965]: DEBUG oslo_vmware.api [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390014, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.443121] env[62965]: DEBUG oslo_vmware.api [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Waiting for the task: (returnval){ [ 915.443121] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52c9b033-5dcf-cec8-413b-cca78ce6d879" [ 915.443121] env[62965]: _type = "Task" [ 915.443121] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.451694] env[62965]: DEBUG oslo_vmware.api [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52c9b033-5dcf-cec8-413b-cca78ce6d879, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.469125] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d6471311-1397-4d5d-875a-0d14109857fb tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.611659] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "0eb943ae-462c-40dd-a870-7a884ea2db74" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.611659] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "0eb943ae-462c-40dd-a870-7a884ea2db74" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.820249] env[62965]: DEBUG oslo_concurrency.lockutils [req-e97cdcfb-a83d-4598-9037-56d444a142c9 req-cdd8865a-30d4-42e5-8e15-cf8faae44b8e service nova] Releasing lock "refresh_cache-00f24976-2f0e-4a2f-8f2e-012b0fce479f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.820249] env[62965]: DEBUG nova.compute.manager [req-e97cdcfb-a83d-4598-9037-56d444a142c9 req-cdd8865a-30d4-42e5-8e15-cf8faae44b8e service nova] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Received event network-vif-deleted-6205e686-0f1c-46bf-a1e7-88f0fd8ee5e5 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 915.820249] env[62965]: DEBUG nova.compute.manager [req-e97cdcfb-a83d-4598-9037-56d444a142c9 req-cdd8865a-30d4-42e5-8e15-cf8faae44b8e service nova] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Received event network-vif-deleted-838a71f4-1d2e-4ffd-8ba7-a09d951c4a66 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 915.851908] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ba31b606-70ea-44b4-80d5-2d58061ebaa5 tempest-InstanceActionsNegativeTestJSON-1319416755 tempest-InstanceActionsNegativeTestJSON-1319416755-project-member] Lock "fa511b61-887e-46a1-aaf6-586d716c0fb5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.196s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.947325] env[62965]: DEBUG oslo_vmware.api [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390015, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.500835} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.947595] env[62965]: DEBUG oslo_vmware.api [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390014, 'name': PowerOnVM_Task, 'duration_secs': 0.761469} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.955354] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] fbbcbe60-92f3-497d-b190-520a2547b3a9/fbbcbe60-92f3-497d-b190-520a2547b3a9.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 915.955601] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 915.955863] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 915.956076] env[62965]: INFO nova.compute.manager [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Took 18.57 seconds to spawn the instance on the hypervisor. [ 915.956292] env[62965]: DEBUG nova.compute.manager [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 915.956964] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-03a98f29-e1f2-4ff1-8613-c6f97c6733a8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.959385] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b59b31e-f389-4d6f-83e8-5b78ee1e5ab9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.968341] env[62965]: DEBUG oslo_vmware.api [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52c9b033-5dcf-cec8-413b-cca78ce6d879, 'name': SearchDatastore_Task, 'duration_secs': 0.009112} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.974059] env[62965]: DEBUG oslo_vmware.api [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 915.974059] env[62965]: value = "task-1390016" [ 915.974059] env[62965]: _type = "Task" [ 915.974059] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.977099] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ecf3ef88-0adf-4b00-973b-7b11eee6bd3f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.986865] env[62965]: DEBUG oslo_vmware.api [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Waiting for the task: (returnval){ [ 915.986865] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52845114-6a2e-ea58-892b-10cc1f99a860" [ 915.986865] env[62965]: _type = "Task" [ 915.986865] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.990107] env[62965]: DEBUG oslo_vmware.api [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390016, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.996239] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a3082e98-a878-445b-a2b1-07e8cce5eb80 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Acquiring lock "64cf898c-2c97-4daa-80c8-0a1c252cbcf2" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.996239] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a3082e98-a878-445b-a2b1-07e8cce5eb80 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Lock "64cf898c-2c97-4daa-80c8-0a1c252cbcf2" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.996472] env[62965]: INFO nova.compute.manager [None req-a3082e98-a878-445b-a2b1-07e8cce5eb80 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Attaching volume c5b95ab6-e0bb-4fcd-a82d-6b0a64749f14 to /dev/sdc [ 916.009405] env[62965]: DEBUG oslo_vmware.api [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52845114-6a2e-ea58-892b-10cc1f99a860, 'name': SearchDatastore_Task, 'duration_secs': 0.009574} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.011769] env[62965]: DEBUG oslo_concurrency.lockutils [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.013134] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 00f24976-2f0e-4a2f-8f2e-012b0fce479f/00f24976-2f0e-4a2f-8f2e-012b0fce479f.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 916.016384] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-71afdf2a-2d61-413e-b19d-c3e54a50d41f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.025162] env[62965]: DEBUG oslo_vmware.api [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Waiting for the task: (returnval){ [ 916.025162] env[62965]: value = "task-1390017" [ 916.025162] env[62965]: _type = "Task" [ 916.025162] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.037807] env[62965]: DEBUG oslo_vmware.api [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390017, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.043051] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-665d7e87-5851-40da-b3f8-09ddd7a72210 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.052726] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0f770c7-a89c-46c2-b731-ab28d9b83312 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.069087] env[62965]: DEBUG nova.virt.block_device [None req-a3082e98-a878-445b-a2b1-07e8cce5eb80 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Updating existing volume attachment record: dc327fa6-ef15-4e27-837e-8d1875c190fa {{(pid=62965) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 916.114461] env[62965]: DEBUG nova.compute.manager [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 916.141744] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d30035de-c2e1-4aa2-a8b4-a094051368d5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.150619] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a517541-bfae-482a-b312-622b0f53a961 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.186274] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a58ab752-807d-4593-b1f6-6cb1611c5363 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.196315] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfcc6c44-3aeb-4d45-81be-221854d2fc53 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.213167] env[62965]: DEBUG nova.compute.provider_tree [None req-48f1ea8b-ebf5-4115-974e-b6c8f24e23b0 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 916.386941] env[62965]: DEBUG nova.compute.manager [None req-1ca81e37-dbeb-4066-88c2-6fba6f88b7d6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 916.388148] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d722976-989e-4138-aa6f-23c385a60c39 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.490344] env[62965]: INFO nova.compute.manager [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Took 27.05 seconds to build instance. [ 916.496728] env[62965]: DEBUG oslo_vmware.api [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390016, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.094165} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.497039] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 916.497833] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b3daf27-8e02-4b7d-94f4-756cb730f007 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.519758] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Reconfiguring VM instance instance-00000052 to attach disk [datastore2] fbbcbe60-92f3-497d-b190-520a2547b3a9/fbbcbe60-92f3-497d-b190-520a2547b3a9.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 916.520632] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-92f5285d-4df0-4c5e-8132-4372cf5b3b01 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.543151] env[62965]: DEBUG oslo_vmware.api [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390017, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.45673} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.544442] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 00f24976-2f0e-4a2f-8f2e-012b0fce479f/00f24976-2f0e-4a2f-8f2e-012b0fce479f.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 916.544664] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 916.544965] env[62965]: DEBUG oslo_vmware.api [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 916.544965] env[62965]: value = "task-1390019" [ 916.544965] env[62965]: _type = "Task" [ 916.544965] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.545173] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f5a00350-50d0-4ce0-9d39-044687909520 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.555881] env[62965]: DEBUG oslo_vmware.api [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390019, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.557084] env[62965]: DEBUG oslo_vmware.api [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Waiting for the task: (returnval){ [ 916.557084] env[62965]: value = "task-1390020" [ 916.557084] env[62965]: _type = "Task" [ 916.557084] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.566399] env[62965]: DEBUG oslo_vmware.api [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390020, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.641117] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.717583] env[62965]: DEBUG nova.scheduler.client.report [None req-48f1ea8b-ebf5-4115-974e-b6c8f24e23b0 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 916.907288] env[62965]: INFO nova.compute.manager [None req-1ca81e37-dbeb-4066-88c2-6fba6f88b7d6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] instance snapshotting [ 916.907288] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-686fd523-4836-4f3b-9c94-e4d10fd73b6a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.933312] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b424d9bd-2b0e-411f-8c47-c3f6d5fe3045 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.992874] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7812458-3e85-4832-8ebd-1647045341e1 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "f82d86ec-3364-449d-be54-35627082c02b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.564s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.057710] env[62965]: DEBUG oslo_vmware.api [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390019, 'name': ReconfigVM_Task, 'duration_secs': 0.295795} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.061174] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Reconfigured VM instance instance-00000052 to attach disk [datastore2] fbbcbe60-92f3-497d-b190-520a2547b3a9/fbbcbe60-92f3-497d-b190-520a2547b3a9.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 917.062226] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6e6a9f58-d6bb-4ce5-aaa7-d13e1a0799a5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.069024] env[62965]: DEBUG oslo_vmware.api [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390020, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064297} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.070846] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 917.071331] env[62965]: DEBUG oslo_vmware.api [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 917.071331] env[62965]: value = "task-1390022" [ 917.071331] env[62965]: _type = "Task" [ 917.071331] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.072185] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95500e70-ef6b-43aa-8961-ef6a2197301a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.083200] env[62965]: DEBUG oslo_vmware.api [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390022, 'name': Rename_Task} progress is 10%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.107205] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Reconfiguring VM instance instance-00000053 to attach disk [datastore2] 00f24976-2f0e-4a2f-8f2e-012b0fce479f/00f24976-2f0e-4a2f-8f2e-012b0fce479f.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 917.107796] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cf7e0052-a9e9-4d46-ac1b-e5ed54a8f441 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.129517] env[62965]: DEBUG oslo_vmware.api [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Waiting for the task: (returnval){ [ 917.129517] env[62965]: value = "task-1390023" [ 917.129517] env[62965]: _type = "Task" [ 917.129517] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.138743] env[62965]: DEBUG oslo_vmware.api [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390023, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.226208] env[62965]: DEBUG oslo_concurrency.lockutils [None req-48f1ea8b-ebf5-4115-974e-b6c8f24e23b0 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.900s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.231532] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d6471311-1397-4d5d-875a-0d14109857fb tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.761s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.231532] env[62965]: DEBUG nova.objects.instance [None req-d6471311-1397-4d5d-875a-0d14109857fb tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lazy-loading 'resources' on Instance uuid eed55d8e-21dc-47d4-a689-415ef6a3891d {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 917.258749] env[62965]: INFO nova.scheduler.client.report [None req-48f1ea8b-ebf5-4115-974e-b6c8f24e23b0 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Deleted allocations for instance 892b1302-f536-4dcd-9227-a545ae1ac731 [ 917.393425] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4a798d04-cb0c-4ce1-925f-11f6078cb448 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquiring lock "f82d86ec-3364-449d-be54-35627082c02b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.393777] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4a798d04-cb0c-4ce1-925f-11f6078cb448 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "f82d86ec-3364-449d-be54-35627082c02b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.394425] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4a798d04-cb0c-4ce1-925f-11f6078cb448 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquiring lock "f82d86ec-3364-449d-be54-35627082c02b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.394425] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4a798d04-cb0c-4ce1-925f-11f6078cb448 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "f82d86ec-3364-449d-be54-35627082c02b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.394425] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4a798d04-cb0c-4ce1-925f-11f6078cb448 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "f82d86ec-3364-449d-be54-35627082c02b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.396626] env[62965]: INFO nova.compute.manager [None req-4a798d04-cb0c-4ce1-925f-11f6078cb448 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Terminating instance [ 917.447082] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-1ca81e37-dbeb-4066-88c2-6fba6f88b7d6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Creating Snapshot of the VM instance {{(pid=62965) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 917.447606] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-035176f8-ca5a-44af-9007-529c98b39ef9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.455924] env[62965]: DEBUG oslo_vmware.api [None req-1ca81e37-dbeb-4066-88c2-6fba6f88b7d6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 917.455924] env[62965]: value = "task-1390024" [ 917.455924] env[62965]: _type = "Task" [ 917.455924] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.466663] env[62965]: DEBUG oslo_vmware.api [None req-1ca81e37-dbeb-4066-88c2-6fba6f88b7d6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390024, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.584192] env[62965]: DEBUG oslo_vmware.api [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390022, 'name': Rename_Task, 'duration_secs': 0.146863} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.584544] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 917.584832] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3d7dffef-14b7-41a7-8286-7caf282f78ff {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.591320] env[62965]: DEBUG oslo_vmware.api [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 917.591320] env[62965]: value = "task-1390025" [ 917.591320] env[62965]: _type = "Task" [ 917.591320] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.598474] env[62965]: DEBUG oslo_vmware.api [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390025, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.640096] env[62965]: DEBUG oslo_vmware.api [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390023, 'name': ReconfigVM_Task, 'duration_secs': 0.298082} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.640430] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Reconfigured VM instance instance-00000053 to attach disk [datastore2] 00f24976-2f0e-4a2f-8f2e-012b0fce479f/00f24976-2f0e-4a2f-8f2e-012b0fce479f.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 917.641164] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a429b0d0-3d3e-4f46-8a26-1e2222b2f95a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.647619] env[62965]: DEBUG oslo_vmware.api [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Waiting for the task: (returnval){ [ 917.647619] env[62965]: value = "task-1390026" [ 917.647619] env[62965]: _type = "Task" [ 917.647619] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.657327] env[62965]: DEBUG oslo_vmware.api [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390026, 'name': Rename_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.770234] env[62965]: DEBUG oslo_concurrency.lockutils [None req-48f1ea8b-ebf5-4115-974e-b6c8f24e23b0 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "892b1302-f536-4dcd-9227-a545ae1ac731" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.744s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.904182] env[62965]: DEBUG nova.compute.manager [None req-4a798d04-cb0c-4ce1-925f-11f6078cb448 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 917.904182] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-4a798d04-cb0c-4ce1-925f-11f6078cb448 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 917.905745] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee6069ac-f856-4a73-b020-00d37660e62b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.915028] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a798d04-cb0c-4ce1-925f-11f6078cb448 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 917.915028] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-571c5219-d6cf-477d-82ec-4231e7935763 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.922025] env[62965]: DEBUG oslo_vmware.api [None req-4a798d04-cb0c-4ce1-925f-11f6078cb448 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 917.922025] env[62965]: value = "task-1390027" [ 917.922025] env[62965]: _type = "Task" [ 917.922025] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.938613] env[62965]: DEBUG oslo_vmware.api [None req-4a798d04-cb0c-4ce1-925f-11f6078cb448 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390027, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.968294] env[62965]: DEBUG oslo_vmware.api [None req-1ca81e37-dbeb-4066-88c2-6fba6f88b7d6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390024, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.981772] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29728148-7363-4561-9c3b-b15c353ae8cd {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.989596] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95537770-f307-4798-a8e2-7592d74f59a7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.021481] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de0e0c9a-0917-436b-9a0f-e60da5701e6c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.029830] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ab3f55c-b38c-48e8-a271-0cf74218fe3a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.043728] env[62965]: DEBUG nova.compute.provider_tree [None req-d6471311-1397-4d5d-875a-0d14109857fb tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 918.101584] env[62965]: DEBUG oslo_vmware.api [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390025, 'name': PowerOnVM_Task, 'duration_secs': 0.476012} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.101584] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 918.101813] env[62965]: INFO nova.compute.manager [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Took 8.78 seconds to spawn the instance on the hypervisor. [ 918.101813] env[62965]: DEBUG nova.compute.manager [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 918.102620] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3342a287-96fb-4b49-8b90-c0fe8c95ab55 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.163248] env[62965]: DEBUG oslo_vmware.api [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390026, 'name': Rename_Task, 'duration_secs': 0.185125} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.163594] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 918.163856] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b171c2ca-32f6-41e8-b7b9-a01a18a7b5b0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.173041] env[62965]: DEBUG oslo_vmware.api [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Waiting for the task: (returnval){ [ 918.173041] env[62965]: value = "task-1390028" [ 918.173041] env[62965]: _type = "Task" [ 918.173041] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.898734] env[62965]: DEBUG nova.scheduler.client.report [None req-d6471311-1397-4d5d-875a-0d14109857fb tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 918.915680] env[62965]: DEBUG oslo_concurrency.lockutils [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "1be808db-d619-4c56-8092-ca52809ec10e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.915942] env[62965]: DEBUG oslo_concurrency.lockutils [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "1be808db-d619-4c56-8092-ca52809ec10e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.917544] env[62965]: INFO nova.compute.manager [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Took 21.27 seconds to build instance. [ 918.923186] env[62965]: DEBUG oslo_vmware.api [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390028, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.923361] env[62965]: WARNING oslo_vmware.common.loopingcall [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] task run outlasted interval by 0.25271200000000005 sec [ 918.933074] env[62965]: DEBUG oslo_vmware.api [None req-4a798d04-cb0c-4ce1-925f-11f6078cb448 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390027, 'name': PowerOffVM_Task, 'duration_secs': 0.320315} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.933530] env[62965]: DEBUG oslo_vmware.api [None req-1ca81e37-dbeb-4066-88c2-6fba6f88b7d6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390024, 'name': CreateSnapshot_Task, 'duration_secs': 1.287847} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.934108] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a798d04-cb0c-4ce1-925f-11f6078cb448 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 918.934293] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-4a798d04-cb0c-4ce1-925f-11f6078cb448 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 918.934548] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-1ca81e37-dbeb-4066-88c2-6fba6f88b7d6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Created Snapshot of the VM instance {{(pid=62965) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 918.934968] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0a7b8a40-6a94-4527-9698-6fa011b6d983 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.938368] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec8d4fdb-950d-4cfa-9c2f-1af7ce66bd02 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.942342] env[62965]: DEBUG oslo_vmware.api [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390028, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.109101] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-4a798d04-cb0c-4ce1-925f-11f6078cb448 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 919.109288] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-4a798d04-cb0c-4ce1-925f-11f6078cb448 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Deleting contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 919.109582] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a798d04-cb0c-4ce1-925f-11f6078cb448 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Deleting the datastore file [datastore1] f82d86ec-3364-449d-be54-35627082c02b {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 919.110413] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-44045191-50bd-4698-9e33-9ca01bdc1e10 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.117660] env[62965]: DEBUG oslo_vmware.api [None req-4a798d04-cb0c-4ce1-925f-11f6078cb448 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 919.117660] env[62965]: value = "task-1390031" [ 919.117660] env[62965]: _type = "Task" [ 919.117660] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.126624] env[62965]: DEBUG oslo_vmware.api [None req-4a798d04-cb0c-4ce1-925f-11f6078cb448 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390031, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.194214] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-624ecc99-34aa-4f6d-b48a-9805c8d1ab8f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Volume attach. Driver type: vmdk {{(pid=62965) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 919.194464] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-624ecc99-34aa-4f6d-b48a-9805c8d1ab8f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-295011', 'volume_id': 'edb73cdb-c23b-4dae-a3b5-8f4c025464d7', 'name': 'volume-edb73cdb-c23b-4dae-a3b5-8f4c025464d7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6def4a38-d3b8-4724-97a1-1859b239d900', 'attached_at': '', 'detached_at': '', 'volume_id': 'edb73cdb-c23b-4dae-a3b5-8f4c025464d7', 'serial': 'edb73cdb-c23b-4dae-a3b5-8f4c025464d7'} {{(pid=62965) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 919.195423] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4a0c59a-3c42-44f8-b3bd-dc4c54ef2e91 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.212108] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b5b3242-b927-46b7-afac-6e61147ddfbe {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.237031] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-624ecc99-34aa-4f6d-b48a-9805c8d1ab8f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] volume-edb73cdb-c23b-4dae-a3b5-8f4c025464d7/volume-edb73cdb-c23b-4dae-a3b5-8f4c025464d7.vmdk or device None with type thin {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 919.237512] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-10c302b5-85c3-4cb3-9a79-f7063385ff52 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.259032] env[62965]: DEBUG oslo_vmware.api [None req-624ecc99-34aa-4f6d-b48a-9805c8d1ab8f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Waiting for the task: (returnval){ [ 919.259032] env[62965]: value = "task-1390032" [ 919.259032] env[62965]: _type = "Task" [ 919.259032] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.267919] env[62965]: DEBUG oslo_vmware.api [None req-624ecc99-34aa-4f6d-b48a-9805c8d1ab8f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390032, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.411296] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d6471311-1397-4d5d-875a-0d14109857fb tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.181s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.413684] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.773s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.415419] env[62965]: INFO nova.compute.claims [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 919.420142] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4ee534c5-228d-4414-8717-55e3dfb23ad0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "fbbcbe60-92f3-497d-b190-520a2547b3a9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.777s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.422728] env[62965]: DEBUG nova.compute.manager [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 919.434029] env[62965]: INFO nova.scheduler.client.report [None req-d6471311-1397-4d5d-875a-0d14109857fb tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Deleted allocations for instance eed55d8e-21dc-47d4-a689-415ef6a3891d [ 919.442062] env[62965]: DEBUG oslo_vmware.api [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390028, 'name': PowerOnVM_Task, 'duration_secs': 1.081822} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.442062] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 919.442062] env[62965]: INFO nova.compute.manager [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Took 7.71 seconds to spawn the instance on the hypervisor. [ 919.442062] env[62965]: DEBUG nova.compute.manager [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 919.442579] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21a4edd8-405e-42d5-88d1-228ad6b18bc0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.458239] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-1ca81e37-dbeb-4066-88c2-6fba6f88b7d6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Creating linked-clone VM from snapshot {{(pid=62965) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 919.458997] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-93d11317-f0ce-4d4a-b80c-ab2072c2faed {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.468713] env[62965]: DEBUG oslo_vmware.api [None req-1ca81e37-dbeb-4066-88c2-6fba6f88b7d6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 919.468713] env[62965]: value = "task-1390033" [ 919.468713] env[62965]: _type = "Task" [ 919.468713] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.477782] env[62965]: DEBUG oslo_vmware.api [None req-1ca81e37-dbeb-4066-88c2-6fba6f88b7d6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390033, 'name': CloneVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.627965] env[62965]: DEBUG oslo_vmware.api [None req-4a798d04-cb0c-4ce1-925f-11f6078cb448 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390031, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149328} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.628271] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a798d04-cb0c-4ce1-925f-11f6078cb448 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 919.628459] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-4a798d04-cb0c-4ce1-925f-11f6078cb448 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Deleted contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 919.629232] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-4a798d04-cb0c-4ce1-925f-11f6078cb448 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 919.629232] env[62965]: INFO nova.compute.manager [None req-4a798d04-cb0c-4ce1-925f-11f6078cb448 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: f82d86ec-3364-449d-be54-35627082c02b] Took 1.73 seconds to destroy the instance on the hypervisor. [ 919.629232] env[62965]: DEBUG oslo.service.loopingcall [None req-4a798d04-cb0c-4ce1-925f-11f6078cb448 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 919.629232] env[62965]: DEBUG nova.compute.manager [-] [instance: f82d86ec-3364-449d-be54-35627082c02b] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 919.629466] env[62965]: DEBUG nova.network.neutron [-] [instance: f82d86ec-3364-449d-be54-35627082c02b] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 919.768229] env[62965]: DEBUG oslo_vmware.api [None req-624ecc99-34aa-4f6d-b48a-9805c8d1ab8f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390032, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.944536] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d6471311-1397-4d5d-875a-0d14109857fb tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lock "eed55d8e-21dc-47d4-a689-415ef6a3891d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.868s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.952176] env[62965]: DEBUG oslo_concurrency.lockutils [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.968805] env[62965]: INFO nova.compute.manager [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Took 20.85 seconds to build instance. [ 919.984521] env[62965]: DEBUG oslo_vmware.api [None req-1ca81e37-dbeb-4066-88c2-6fba6f88b7d6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390033, 'name': CloneVM_Task} progress is 94%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.059429] env[62965]: DEBUG nova.compute.manager [req-13e036ad-8405-4bb1-b73d-7597ccd71528 req-2d94435b-e882-4dda-befb-ce21b19ec547 service nova] [instance: f82d86ec-3364-449d-be54-35627082c02b] Received event network-vif-deleted-ff5bbb06-f9d8-4d9c-8884-d85b3dbbb90e {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 920.059709] env[62965]: INFO nova.compute.manager [req-13e036ad-8405-4bb1-b73d-7597ccd71528 req-2d94435b-e882-4dda-befb-ce21b19ec547 service nova] [instance: f82d86ec-3364-449d-be54-35627082c02b] Neutron deleted interface ff5bbb06-f9d8-4d9c-8884-d85b3dbbb90e; detaching it from the instance and deleting it from the info cache [ 920.059938] env[62965]: DEBUG nova.network.neutron [req-13e036ad-8405-4bb1-b73d-7597ccd71528 req-2d94435b-e882-4dda-befb-ce21b19ec547 service nova] [instance: f82d86ec-3364-449d-be54-35627082c02b] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.274960] env[62965]: DEBUG oslo_vmware.api [None req-624ecc99-34aa-4f6d-b48a-9805c8d1ab8f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390032, 'name': ReconfigVM_Task, 'duration_secs': 0.763598} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.275470] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-624ecc99-34aa-4f6d-b48a-9805c8d1ab8f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Reconfigured VM instance instance-00000047 to attach disk [datastore2] volume-edb73cdb-c23b-4dae-a3b5-8f4c025464d7/volume-edb73cdb-c23b-4dae-a3b5-8f4c025464d7.vmdk or device None with type thin {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 920.281802] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7f75d033-f63b-4b6d-8ddb-d63245fd166f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.300644] env[62965]: DEBUG oslo_vmware.api [None req-624ecc99-34aa-4f6d-b48a-9805c8d1ab8f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Waiting for the task: (returnval){ [ 920.300644] env[62965]: value = "task-1390034" [ 920.300644] env[62965]: _type = "Task" [ 920.300644] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.309664] env[62965]: DEBUG oslo_vmware.api [None req-624ecc99-34aa-4f6d-b48a-9805c8d1ab8f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390034, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.311290] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquiring lock "b59d6023-a047-4b32-bd18-1c7cc05f67d7" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.311666] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "b59d6023-a047-4b32-bd18-1c7cc05f67d7" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.313207] env[62965]: INFO nova.compute.manager [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Shelving [ 920.341324] env[62965]: DEBUG nova.compute.manager [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Stashing vm_state: active {{(pid=62965) _prep_resize /opt/stack/nova/nova/compute/manager.py:5953}} [ 920.414130] env[62965]: DEBUG nova.network.neutron [-] [instance: f82d86ec-3364-449d-be54-35627082c02b] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.483032] env[62965]: DEBUG oslo_concurrency.lockutils [None req-55b5ecdb-c9a1-4e86-a5f7-53e0c9e88df5 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "00f24976-2f0e-4a2f-8f2e-012b0fce479f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.376s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.483927] env[62965]: DEBUG oslo_vmware.api [None req-1ca81e37-dbeb-4066-88c2-6fba6f88b7d6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390033, 'name': CloneVM_Task} progress is 94%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.564965] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0126f68b-6972-4b67-90ad-f2b8b4d38c1c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.573959] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b7d5b1d-3ab1-4d3f-93be-f58685215e92 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.608743] env[62965]: DEBUG nova.compute.manager [req-13e036ad-8405-4bb1-b73d-7597ccd71528 req-2d94435b-e882-4dda-befb-ce21b19ec547 service nova] [instance: f82d86ec-3364-449d-be54-35627082c02b] Detach interface failed, port_id=ff5bbb06-f9d8-4d9c-8884-d85b3dbbb90e, reason: Instance f82d86ec-3364-449d-be54-35627082c02b could not be found. {{(pid=62965) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11294}} [ 920.624036] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3082e98-a878-445b-a2b1-07e8cce5eb80 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Volume attach. Driver type: vmdk {{(pid=62965) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 920.624316] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3082e98-a878-445b-a2b1-07e8cce5eb80 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-295012', 'volume_id': 'c5b95ab6-e0bb-4fcd-a82d-6b0a64749f14', 'name': 'volume-c5b95ab6-e0bb-4fcd-a82d-6b0a64749f14', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '64cf898c-2c97-4daa-80c8-0a1c252cbcf2', 'attached_at': '', 'detached_at': '', 'volume_id': 'c5b95ab6-e0bb-4fcd-a82d-6b0a64749f14', 'serial': 'c5b95ab6-e0bb-4fcd-a82d-6b0a64749f14'} {{(pid=62965) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 920.625273] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-221e3c07-a524-4507-8334-02c618836761 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.648747] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d057f26-6f73-49e5-9dfd-46685a03e8bd {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.677939] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3082e98-a878-445b-a2b1-07e8cce5eb80 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] volume-c5b95ab6-e0bb-4fcd-a82d-6b0a64749f14/volume-c5b95ab6-e0bb-4fcd-a82d-6b0a64749f14.vmdk or device None with type thin {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 920.680855] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fd278d75-b5f4-4829-92ab-aa8b2d924470 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.699982] env[62965]: DEBUG oslo_vmware.api [None req-a3082e98-a878-445b-a2b1-07e8cce5eb80 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Waiting for the task: (returnval){ [ 920.699982] env[62965]: value = "task-1390035" [ 920.699982] env[62965]: _type = "Task" [ 920.699982] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.711249] env[62965]: DEBUG oslo_vmware.api [None req-a3082e98-a878-445b-a2b1-07e8cce5eb80 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1390035, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.742925] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6913ed44-454c-4fb7-a215-473302bc54d1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.750821] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-307a8d5b-d07c-4b67-844a-278aa76ce989 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.782439] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d15e3e93-1b01-4edc-a18b-e2678813aaaa {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.791014] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aedf1cf-35f8-4838-bdd7-17337da1881d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.804570] env[62965]: DEBUG nova.compute.provider_tree [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 920.813663] env[62965]: DEBUG oslo_vmware.api [None req-624ecc99-34aa-4f6d-b48a-9805c8d1ab8f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390034, 'name': ReconfigVM_Task, 'duration_secs': 0.160584} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.813972] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-624ecc99-34aa-4f6d-b48a-9805c8d1ab8f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-295011', 'volume_id': 'edb73cdb-c23b-4dae-a3b5-8f4c025464d7', 'name': 'volume-edb73cdb-c23b-4dae-a3b5-8f4c025464d7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6def4a38-d3b8-4724-97a1-1859b239d900', 'attached_at': '', 'detached_at': '', 'volume_id': 'edb73cdb-c23b-4dae-a3b5-8f4c025464d7', 'serial': 'edb73cdb-c23b-4dae-a3b5-8f4c025464d7'} {{(pid=62965) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 920.860399] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.917364] env[62965]: INFO nova.compute.manager [-] [instance: f82d86ec-3364-449d-be54-35627082c02b] Took 1.29 seconds to deallocate network for instance. [ 920.981934] env[62965]: DEBUG oslo_vmware.api [None req-1ca81e37-dbeb-4066-88c2-6fba6f88b7d6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390033, 'name': CloneVM_Task, 'duration_secs': 1.38999} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.982298] env[62965]: INFO nova.virt.vmwareapi.vmops [None req-1ca81e37-dbeb-4066-88c2-6fba6f88b7d6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Created linked-clone VM from snapshot [ 920.982954] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b8d3f2b-b555-457d-a9ef-53d8920edc77 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.991181] env[62965]: DEBUG nova.virt.vmwareapi.images [None req-1ca81e37-dbeb-4066-88c2-6fba6f88b7d6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Uploading image 28513290-2c05-4b98-aaab-c1ce6da167df {{(pid=62965) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 921.014890] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ca81e37-dbeb-4066-88c2-6fba6f88b7d6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Destroying the VM {{(pid=62965) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 921.015256] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-9eafb80f-c0a4-435f-bf23-c4f700480d84 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.026367] env[62965]: DEBUG oslo_vmware.api [None req-1ca81e37-dbeb-4066-88c2-6fba6f88b7d6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 921.026367] env[62965]: value = "task-1390036" [ 921.026367] env[62965]: _type = "Task" [ 921.026367] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.036325] env[62965]: DEBUG oslo_vmware.api [None req-1ca81e37-dbeb-4066-88c2-6fba6f88b7d6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390036, 'name': Destroy_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.210749] env[62965]: DEBUG oslo_vmware.api [None req-a3082e98-a878-445b-a2b1-07e8cce5eb80 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1390035, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.311023] env[62965]: DEBUG nova.scheduler.client.report [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 921.323736] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 921.324057] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-158279c6-7e2b-4fc9-b584-709867b96f05 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.332624] env[62965]: DEBUG oslo_vmware.api [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 921.332624] env[62965]: value = "task-1390037" [ 921.332624] env[62965]: _type = "Task" [ 921.332624] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.340837] env[62965]: DEBUG oslo_vmware.api [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390037, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.425971] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4a798d04-cb0c-4ce1-925f-11f6078cb448 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.536827] env[62965]: DEBUG oslo_vmware.api [None req-1ca81e37-dbeb-4066-88c2-6fba6f88b7d6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390036, 'name': Destroy_Task} progress is 33%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.712133] env[62965]: DEBUG oslo_vmware.api [None req-a3082e98-a878-445b-a2b1-07e8cce5eb80 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1390035, 'name': ReconfigVM_Task, 'duration_secs': 0.591796} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.712547] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3082e98-a878-445b-a2b1-07e8cce5eb80 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Reconfigured VM instance instance-00000040 to attach disk [datastore2] volume-c5b95ab6-e0bb-4fcd-a82d-6b0a64749f14/volume-c5b95ab6-e0bb-4fcd-a82d-6b0a64749f14.vmdk or device None with type thin {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 921.717987] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-66887db5-958d-4bf0-af95-883901d0d5a7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.734291] env[62965]: DEBUG oslo_vmware.api [None req-a3082e98-a878-445b-a2b1-07e8cce5eb80 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Waiting for the task: (returnval){ [ 921.734291] env[62965]: value = "task-1390038" [ 921.734291] env[62965]: _type = "Task" [ 921.734291] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.743913] env[62965]: DEBUG oslo_vmware.api [None req-a3082e98-a878-445b-a2b1-07e8cce5eb80 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1390038, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.816124] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.402s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.816826] env[62965]: DEBUG nova.compute.manager [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 921.819803] env[62965]: DEBUG oslo_concurrency.lockutils [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.868s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.821500] env[62965]: INFO nova.compute.claims [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 921.844520] env[62965]: DEBUG oslo_vmware.api [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390037, 'name': PowerOffVM_Task, 'duration_secs': 0.247414} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.844520] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 921.845054] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-950bdb9b-3e5d-423a-a475-38f86c1ff7a9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.867551] env[62965]: DEBUG nova.objects.instance [None req-624ecc99-34aa-4f6d-b48a-9805c8d1ab8f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lazy-loading 'flavor' on Instance uuid 6def4a38-d3b8-4724-97a1-1859b239d900 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 921.871159] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-413f52d3-3211-4b59-948f-2be9c4160e6a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.038447] env[62965]: DEBUG oslo_vmware.api [None req-1ca81e37-dbeb-4066-88c2-6fba6f88b7d6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390036, 'name': Destroy_Task, 'duration_secs': 0.751618} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.038845] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-1ca81e37-dbeb-4066-88c2-6fba6f88b7d6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Destroyed the VM [ 922.038991] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-1ca81e37-dbeb-4066-88c2-6fba6f88b7d6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Deleting Snapshot of the VM instance {{(pid=62965) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 922.039290] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-9ee99798-b3a2-4aae-a9b2-9f572eda24b3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.046638] env[62965]: DEBUG oslo_vmware.api [None req-1ca81e37-dbeb-4066-88c2-6fba6f88b7d6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 922.046638] env[62965]: value = "task-1390039" [ 922.046638] env[62965]: _type = "Task" [ 922.046638] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.054568] env[62965]: DEBUG oslo_vmware.api [None req-1ca81e37-dbeb-4066-88c2-6fba6f88b7d6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390039, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.099325] env[62965]: DEBUG nova.compute.manager [req-2c42f935-b0e3-4178-89da-beb8f378b5ff req-ebce7590-8c2f-4d34-a1b5-7daa8e05a90d service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Received event network-changed-b1718061-edf1-4077-b57f-0e54aea365a6 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 922.099522] env[62965]: DEBUG nova.compute.manager [req-2c42f935-b0e3-4178-89da-beb8f378b5ff req-ebce7590-8c2f-4d34-a1b5-7daa8e05a90d service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Refreshing instance network info cache due to event network-changed-b1718061-edf1-4077-b57f-0e54aea365a6. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 922.100129] env[62965]: DEBUG oslo_concurrency.lockutils [req-2c42f935-b0e3-4178-89da-beb8f378b5ff req-ebce7590-8c2f-4d34-a1b5-7daa8e05a90d service nova] Acquiring lock "refresh_cache-d038608c-9264-4fd5-b9ba-0653c98ec0e4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.100387] env[62965]: DEBUG oslo_concurrency.lockutils [req-2c42f935-b0e3-4178-89da-beb8f378b5ff req-ebce7590-8c2f-4d34-a1b5-7daa8e05a90d service nova] Acquired lock "refresh_cache-d038608c-9264-4fd5-b9ba-0653c98ec0e4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.100471] env[62965]: DEBUG nova.network.neutron [req-2c42f935-b0e3-4178-89da-beb8f378b5ff req-ebce7590-8c2f-4d34-a1b5-7daa8e05a90d service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Refreshing network info cache for port b1718061-edf1-4077-b57f-0e54aea365a6 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 922.151721] env[62965]: DEBUG nova.compute.manager [req-119fd488-404d-4a5a-9dbb-d80931a3ce59 req-a868c5ac-8df9-4c5a-b9c9-07f328b240b4 service nova] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Received event network-changed-8fea9d40-d98f-4328-ad3e-ae708a849908 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 922.151870] env[62965]: DEBUG nova.compute.manager [req-119fd488-404d-4a5a-9dbb-d80931a3ce59 req-a868c5ac-8df9-4c5a-b9c9-07f328b240b4 service nova] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Refreshing instance network info cache due to event network-changed-8fea9d40-d98f-4328-ad3e-ae708a849908. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 922.152422] env[62965]: DEBUG oslo_concurrency.lockutils [req-119fd488-404d-4a5a-9dbb-d80931a3ce59 req-a868c5ac-8df9-4c5a-b9c9-07f328b240b4 service nova] Acquiring lock "refresh_cache-00f24976-2f0e-4a2f-8f2e-012b0fce479f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.152597] env[62965]: DEBUG oslo_concurrency.lockutils [req-119fd488-404d-4a5a-9dbb-d80931a3ce59 req-a868c5ac-8df9-4c5a-b9c9-07f328b240b4 service nova] Acquired lock "refresh_cache-00f24976-2f0e-4a2f-8f2e-012b0fce479f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.152804] env[62965]: DEBUG nova.network.neutron [req-119fd488-404d-4a5a-9dbb-d80931a3ce59 req-a868c5ac-8df9-4c5a-b9c9-07f328b240b4 service nova] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Refreshing network info cache for port 8fea9d40-d98f-4328-ad3e-ae708a849908 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 922.250356] env[62965]: DEBUG oslo_vmware.api [None req-a3082e98-a878-445b-a2b1-07e8cce5eb80 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1390038, 'name': ReconfigVM_Task, 'duration_secs': 0.179871} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.250720] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-a3082e98-a878-445b-a2b1-07e8cce5eb80 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-295012', 'volume_id': 'c5b95ab6-e0bb-4fcd-a82d-6b0a64749f14', 'name': 'volume-c5b95ab6-e0bb-4fcd-a82d-6b0a64749f14', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '64cf898c-2c97-4daa-80c8-0a1c252cbcf2', 'attached_at': '', 'detached_at': '', 'volume_id': 'c5b95ab6-e0bb-4fcd-a82d-6b0a64749f14', 'serial': 'c5b95ab6-e0bb-4fcd-a82d-6b0a64749f14'} {{(pid=62965) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 922.326404] env[62965]: DEBUG nova.compute.utils [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 922.330790] env[62965]: DEBUG nova.compute.manager [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 922.330790] env[62965]: DEBUG nova.network.neutron [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 922.371601] env[62965]: DEBUG nova.policy [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03c0a9a0ec3c46bf883c68d86cb816d6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '978666785b0743bb843de96585f96fb4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 922.377111] env[62965]: DEBUG oslo_concurrency.lockutils [None req-624ecc99-34aa-4f6d-b48a-9805c8d1ab8f tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "6def4a38-d3b8-4724-97a1-1859b239d900" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.789s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.384255] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Creating Snapshot of the VM instance {{(pid=62965) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 922.384598] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-578822c3-e93b-47eb-bc55-2b65a950f846 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.394138] env[62965]: DEBUG oslo_vmware.api [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 922.394138] env[62965]: value = "task-1390040" [ 922.394138] env[62965]: _type = "Task" [ 922.394138] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.402742] env[62965]: DEBUG oslo_vmware.api [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390040, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.557345] env[62965]: DEBUG oslo_vmware.api [None req-1ca81e37-dbeb-4066-88c2-6fba6f88b7d6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390039, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.753258] env[62965]: DEBUG nova.network.neutron [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Successfully created port: c6f37c93-3ad8-4408-9f81-519246ae7051 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 922.842052] env[62965]: DEBUG nova.compute.manager [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 922.908381] env[62965]: DEBUG oslo_vmware.api [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390040, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.060109] env[62965]: DEBUG oslo_vmware.api [None req-1ca81e37-dbeb-4066-88c2-6fba6f88b7d6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390039, 'name': RemoveSnapshot_Task, 'duration_secs': 0.825405} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.060486] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-1ca81e37-dbeb-4066-88c2-6fba6f88b7d6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Deleted Snapshot of the VM instance {{(pid=62965) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 923.134020] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fa3c5ba-9fb3-48c6-811e-4bc95e038cac {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.138859] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c57efebf-469b-44e1-aaea-2baf5b00df9a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.143442] env[62965]: DEBUG oslo_concurrency.lockutils [None req-36b4d40d-b72a-4a0e-a875-67e53da32f82 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquiring lock "6def4a38-d3b8-4724-97a1-1859b239d900" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.143752] env[62965]: DEBUG oslo_concurrency.lockutils [None req-36b4d40d-b72a-4a0e-a875-67e53da32f82 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "6def4a38-d3b8-4724-97a1-1859b239d900" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.173400] env[62965]: INFO nova.compute.manager [None req-36b4d40d-b72a-4a0e-a875-67e53da32f82 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Detaching volume edb73cdb-c23b-4dae-a3b5-8f4c025464d7 [ 923.175683] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2710f3a6-d565-4ecf-9a21-fc7b60666f5f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.184642] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f849515e-2ee6-4c0b-9b58-92666808f792 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.204136] env[62965]: DEBUG nova.compute.provider_tree [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 923.210082] env[62965]: DEBUG nova.network.neutron [req-119fd488-404d-4a5a-9dbb-d80931a3ce59 req-a868c5ac-8df9-4c5a-b9c9-07f328b240b4 service nova] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Updated VIF entry in instance network info cache for port 8fea9d40-d98f-4328-ad3e-ae708a849908. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 923.210082] env[62965]: DEBUG nova.network.neutron [req-119fd488-404d-4a5a-9dbb-d80931a3ce59 req-a868c5ac-8df9-4c5a-b9c9-07f328b240b4 service nova] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Updating instance_info_cache with network_info: [{"id": "8fea9d40-d98f-4328-ad3e-ae708a849908", "address": "fa:16:3e:c3:8c:35", "network": {"id": "a9699d8b-5e76-4666-a236-33ab8f22345e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1329540850-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1952dedf340f4b7ba0d57af6bba9a749", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a316376e-2ef0-4b1e-b40c-10321ebd7e1a", "external-id": "nsx-vlan-transportzone-942", "segmentation_id": 942, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8fea9d40-d9", "ovs_interfaceid": "8fea9d40-d98f-4328-ad3e-ae708a849908", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.213943] env[62965]: DEBUG nova.network.neutron [req-2c42f935-b0e3-4178-89da-beb8f378b5ff req-ebce7590-8c2f-4d34-a1b5-7daa8e05a90d service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Updated VIF entry in instance network info cache for port b1718061-edf1-4077-b57f-0e54aea365a6. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 923.214296] env[62965]: DEBUG nova.network.neutron [req-2c42f935-b0e3-4178-89da-beb8f378b5ff req-ebce7590-8c2f-4d34-a1b5-7daa8e05a90d service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Updating instance_info_cache with network_info: [{"id": "b1718061-edf1-4077-b57f-0e54aea365a6", "address": "fa:16:3e:1f:27:16", "network": {"id": "a9699d8b-5e76-4666-a236-33ab8f22345e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1329540850-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1952dedf340f4b7ba0d57af6bba9a749", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a316376e-2ef0-4b1e-b40c-10321ebd7e1a", "external-id": "nsx-vlan-transportzone-942", "segmentation_id": 942, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1718061-ed", "ovs_interfaceid": "b1718061-edf1-4077-b57f-0e54aea365a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.231117] env[62965]: INFO nova.virt.block_device [None req-36b4d40d-b72a-4a0e-a875-67e53da32f82 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Attempting to driver detach volume edb73cdb-c23b-4dae-a3b5-8f4c025464d7 from mountpoint /dev/sdb [ 923.231240] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-36b4d40d-b72a-4a0e-a875-67e53da32f82 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Volume detach. Driver type: vmdk {{(pid=62965) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 923.231482] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-36b4d40d-b72a-4a0e-a875-67e53da32f82 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-295011', 'volume_id': 'edb73cdb-c23b-4dae-a3b5-8f4c025464d7', 'name': 'volume-edb73cdb-c23b-4dae-a3b5-8f4c025464d7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6def4a38-d3b8-4724-97a1-1859b239d900', 'attached_at': '', 'detached_at': '', 'volume_id': 'edb73cdb-c23b-4dae-a3b5-8f4c025464d7', 'serial': 'edb73cdb-c23b-4dae-a3b5-8f4c025464d7'} {{(pid=62965) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 923.235693] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3417d53e-3949-47a1-ad94-1c353ce48ef8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.261254] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9469957-6d97-4c82-b53b-af6b4ec0c3c0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.268964] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16ecb673-1f55-40e2-a88c-d000357d51a2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.296810] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88bbfdfe-fcaa-42f7-8a0e-27c400365c06 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.309809] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-36b4d40d-b72a-4a0e-a875-67e53da32f82 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] The volume has not been displaced from its original location: [datastore2] volume-edb73cdb-c23b-4dae-a3b5-8f4c025464d7/volume-edb73cdb-c23b-4dae-a3b5-8f4c025464d7.vmdk. No consolidation needed. {{(pid=62965) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 923.316303] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-36b4d40d-b72a-4a0e-a875-67e53da32f82 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Reconfiguring VM instance instance-00000047 to detach disk 2001 {{(pid=62965) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 923.317105] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c4a0cb54-0a26-4e6f-ae0b-aa15c054b61c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.340525] env[62965]: DEBUG oslo_vmware.api [None req-36b4d40d-b72a-4a0e-a875-67e53da32f82 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Waiting for the task: (returnval){ [ 923.340525] env[62965]: value = "task-1390041" [ 923.340525] env[62965]: _type = "Task" [ 923.340525] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.349401] env[62965]: DEBUG nova.objects.instance [None req-a3082e98-a878-445b-a2b1-07e8cce5eb80 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Lazy-loading 'flavor' on Instance uuid 64cf898c-2c97-4daa-80c8-0a1c252cbcf2 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 923.354166] env[62965]: DEBUG oslo_vmware.api [None req-36b4d40d-b72a-4a0e-a875-67e53da32f82 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390041, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.408765] env[62965]: DEBUG oslo_vmware.api [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390040, 'name': CreateSnapshot_Task, 'duration_secs': 0.759671} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.408765] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Created Snapshot of the VM instance {{(pid=62965) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 923.409169] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46ce994a-4075-44eb-a04c-ba479fd8f576 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.564935] env[62965]: WARNING nova.compute.manager [None req-1ca81e37-dbeb-4066-88c2-6fba6f88b7d6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Image not found during snapshot: nova.exception.ImageNotFound: Image 28513290-2c05-4b98-aaab-c1ce6da167df could not be found. [ 923.709158] env[62965]: DEBUG nova.scheduler.client.report [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 923.713098] env[62965]: DEBUG oslo_concurrency.lockutils [req-119fd488-404d-4a5a-9dbb-d80931a3ce59 req-a868c5ac-8df9-4c5a-b9c9-07f328b240b4 service nova] Releasing lock "refresh_cache-00f24976-2f0e-4a2f-8f2e-012b0fce479f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.720202] env[62965]: DEBUG oslo_concurrency.lockutils [req-2c42f935-b0e3-4178-89da-beb8f378b5ff req-ebce7590-8c2f-4d34-a1b5-7daa8e05a90d service nova] Releasing lock "refresh_cache-d038608c-9264-4fd5-b9ba-0653c98ec0e4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.720202] env[62965]: DEBUG nova.compute.manager [req-2c42f935-b0e3-4178-89da-beb8f378b5ff req-ebce7590-8c2f-4d34-a1b5-7daa8e05a90d service nova] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Received event network-changed-8fea9d40-d98f-4328-ad3e-ae708a849908 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 923.720382] env[62965]: DEBUG nova.compute.manager [req-2c42f935-b0e3-4178-89da-beb8f378b5ff req-ebce7590-8c2f-4d34-a1b5-7daa8e05a90d service nova] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Refreshing instance network info cache due to event network-changed-8fea9d40-d98f-4328-ad3e-ae708a849908. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 923.720597] env[62965]: DEBUG oslo_concurrency.lockutils [req-2c42f935-b0e3-4178-89da-beb8f378b5ff req-ebce7590-8c2f-4d34-a1b5-7daa8e05a90d service nova] Acquiring lock "refresh_cache-00f24976-2f0e-4a2f-8f2e-012b0fce479f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.720742] env[62965]: DEBUG oslo_concurrency.lockutils [req-2c42f935-b0e3-4178-89da-beb8f378b5ff req-ebce7590-8c2f-4d34-a1b5-7daa8e05a90d service nova] Acquired lock "refresh_cache-00f24976-2f0e-4a2f-8f2e-012b0fce479f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.720923] env[62965]: DEBUG nova.network.neutron [req-2c42f935-b0e3-4178-89da-beb8f378b5ff req-ebce7590-8c2f-4d34-a1b5-7daa8e05a90d service nova] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Refreshing network info cache for port 8fea9d40-d98f-4328-ad3e-ae708a849908 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 923.835519] env[62965]: DEBUG oslo_concurrency.lockutils [None req-23c6bb8c-957a-4b64-9cee-aa35e0118c0a tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Acquiring lock "e0c4e188-a041-4aa7-95de-6e8718b17ade" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.835798] env[62965]: DEBUG oslo_concurrency.lockutils [None req-23c6bb8c-957a-4b64-9cee-aa35e0118c0a tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Lock "e0c4e188-a041-4aa7-95de-6e8718b17ade" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.836056] env[62965]: DEBUG oslo_concurrency.lockutils [None req-23c6bb8c-957a-4b64-9cee-aa35e0118c0a tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Acquiring lock "e0c4e188-a041-4aa7-95de-6e8718b17ade-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.836272] env[62965]: DEBUG oslo_concurrency.lockutils [None req-23c6bb8c-957a-4b64-9cee-aa35e0118c0a tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Lock "e0c4e188-a041-4aa7-95de-6e8718b17ade-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.836461] env[62965]: DEBUG oslo_concurrency.lockutils [None req-23c6bb8c-957a-4b64-9cee-aa35e0118c0a tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Lock "e0c4e188-a041-4aa7-95de-6e8718b17ade-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.838417] env[62965]: INFO nova.compute.manager [None req-23c6bb8c-957a-4b64-9cee-aa35e0118c0a tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Terminating instance [ 923.853163] env[62965]: DEBUG nova.compute.manager [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 923.855336] env[62965]: DEBUG oslo_vmware.api [None req-36b4d40d-b72a-4a0e-a875-67e53da32f82 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390041, 'name': ReconfigVM_Task, 'duration_secs': 0.271037} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.857203] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-36b4d40d-b72a-4a0e-a875-67e53da32f82 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Reconfigured VM instance instance-00000047 to detach disk 2001 {{(pid=62965) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 923.862202] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a3082e98-a878-445b-a2b1-07e8cce5eb80 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Lock "64cf898c-2c97-4daa-80c8-0a1c252cbcf2" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.866s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.864039] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3bb138f4-698a-4db1-b091-6990621ca3d2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.880465] env[62965]: DEBUG oslo_vmware.api [None req-36b4d40d-b72a-4a0e-a875-67e53da32f82 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Waiting for the task: (returnval){ [ 923.880465] env[62965]: value = "task-1390042" [ 923.880465] env[62965]: _type = "Task" [ 923.880465] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.882480] env[62965]: DEBUG nova.virt.hardware [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 923.882704] env[62965]: DEBUG nova.virt.hardware [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 923.882862] env[62965]: DEBUG nova.virt.hardware [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 923.883061] env[62965]: DEBUG nova.virt.hardware [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 923.883214] env[62965]: DEBUG nova.virt.hardware [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 923.883363] env[62965]: DEBUG nova.virt.hardware [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 923.883566] env[62965]: DEBUG nova.virt.hardware [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 923.883727] env[62965]: DEBUG nova.virt.hardware [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 923.883892] env[62965]: DEBUG nova.virt.hardware [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 923.884074] env[62965]: DEBUG nova.virt.hardware [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 923.884256] env[62965]: DEBUG nova.virt.hardware [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 923.885269] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47c36bb7-bc4c-4777-bc8d-8c92f5dd9757 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.897093] env[62965]: DEBUG oslo_vmware.api [None req-36b4d40d-b72a-4a0e-a875-67e53da32f82 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390042, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.899655] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c3940c0-17e4-4db3-8270-abb8e545fb59 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.926521] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Creating linked-clone VM from snapshot {{(pid=62965) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 923.926997] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-35f72167-6a34-432f-b448-da9ff159b309 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.934900] env[62965]: DEBUG oslo_vmware.api [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 923.934900] env[62965]: value = "task-1390043" [ 923.934900] env[62965]: _type = "Task" [ 923.934900] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.942611] env[62965]: DEBUG oslo_vmware.api [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390043, 'name': CloneVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.176385] env[62965]: DEBUG nova.compute.manager [req-578e0ad4-7ef2-4b33-90d1-71ae5e4e5c64 req-13f4f758-f244-484b-8e53-95186ce60a04 service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Received event network-changed-b1718061-edf1-4077-b57f-0e54aea365a6 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 924.176617] env[62965]: DEBUG nova.compute.manager [req-578e0ad4-7ef2-4b33-90d1-71ae5e4e5c64 req-13f4f758-f244-484b-8e53-95186ce60a04 service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Refreshing instance network info cache due to event network-changed-b1718061-edf1-4077-b57f-0e54aea365a6. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 924.176765] env[62965]: DEBUG oslo_concurrency.lockutils [req-578e0ad4-7ef2-4b33-90d1-71ae5e4e5c64 req-13f4f758-f244-484b-8e53-95186ce60a04 service nova] Acquiring lock "refresh_cache-d038608c-9264-4fd5-b9ba-0653c98ec0e4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.176912] env[62965]: DEBUG oslo_concurrency.lockutils [req-578e0ad4-7ef2-4b33-90d1-71ae5e4e5c64 req-13f4f758-f244-484b-8e53-95186ce60a04 service nova] Acquired lock "refresh_cache-d038608c-9264-4fd5-b9ba-0653c98ec0e4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.177124] env[62965]: DEBUG nova.network.neutron [req-578e0ad4-7ef2-4b33-90d1-71ae5e4e5c64 req-13f4f758-f244-484b-8e53-95186ce60a04 service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Refreshing network info cache for port b1718061-edf1-4077-b57f-0e54aea365a6 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 924.214421] env[62965]: DEBUG oslo_concurrency.lockutils [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.395s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.214950] env[62965]: DEBUG nova.compute.manager [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 924.219782] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 3.360s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.320365] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ce981bd3-6c64-4571-885c-7e3dd74e2d52 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Acquiring lock "64cf898c-2c97-4daa-80c8-0a1c252cbcf2" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.320599] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ce981bd3-6c64-4571-885c-7e3dd74e2d52 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Lock "64cf898c-2c97-4daa-80c8-0a1c252cbcf2" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.341754] env[62965]: DEBUG nova.compute.manager [None req-23c6bb8c-957a-4b64-9cee-aa35e0118c0a tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 924.341971] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-23c6bb8c-957a-4b64-9cee-aa35e0118c0a tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 924.343523] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33264585-a6c0-4d7a-a048-7ef3e5597255 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.353304] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-23c6bb8c-957a-4b64-9cee-aa35e0118c0a tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 924.353561] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fa7b8056-3ad5-4ef4-8bce-144a12127548 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.359479] env[62965]: DEBUG oslo_vmware.api [None req-23c6bb8c-957a-4b64-9cee-aa35e0118c0a tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 924.359479] env[62965]: value = "task-1390044" [ 924.359479] env[62965]: _type = "Task" [ 924.359479] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.367326] env[62965]: DEBUG oslo_vmware.api [None req-23c6bb8c-957a-4b64-9cee-aa35e0118c0a tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390044, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.397031] env[62965]: DEBUG oslo_vmware.api [None req-36b4d40d-b72a-4a0e-a875-67e53da32f82 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390042, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.444797] env[62965]: DEBUG oslo_vmware.api [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390043, 'name': CloneVM_Task} progress is 94%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.502606] env[62965]: DEBUG nova.network.neutron [req-2c42f935-b0e3-4178-89da-beb8f378b5ff req-ebce7590-8c2f-4d34-a1b5-7daa8e05a90d service nova] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Updated VIF entry in instance network info cache for port 8fea9d40-d98f-4328-ad3e-ae708a849908. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 924.503165] env[62965]: DEBUG nova.network.neutron [req-2c42f935-b0e3-4178-89da-beb8f378b5ff req-ebce7590-8c2f-4d34-a1b5-7daa8e05a90d service nova] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Updating instance_info_cache with network_info: [{"id": "8fea9d40-d98f-4328-ad3e-ae708a849908", "address": "fa:16:3e:c3:8c:35", "network": {"id": "a9699d8b-5e76-4666-a236-33ab8f22345e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1329540850-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1952dedf340f4b7ba0d57af6bba9a749", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a316376e-2ef0-4b1e-b40c-10321ebd7e1a", "external-id": "nsx-vlan-transportzone-942", "segmentation_id": 942, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8fea9d40-d9", "ovs_interfaceid": "8fea9d40-d98f-4328-ad3e-ae708a849908", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.630908] env[62965]: DEBUG nova.compute.manager [req-4ca6a2d1-753f-426a-b542-cfc0f777c991 req-6f38726d-1b11-4012-844f-38c0fb6ecd7b service nova] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Received event network-vif-plugged-c6f37c93-3ad8-4408-9f81-519246ae7051 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 924.633459] env[62965]: DEBUG oslo_concurrency.lockutils [req-4ca6a2d1-753f-426a-b542-cfc0f777c991 req-6f38726d-1b11-4012-844f-38c0fb6ecd7b service nova] Acquiring lock "0eb943ae-462c-40dd-a870-7a884ea2db74-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.633714] env[62965]: DEBUG oslo_concurrency.lockutils [req-4ca6a2d1-753f-426a-b542-cfc0f777c991 req-6f38726d-1b11-4012-844f-38c0fb6ecd7b service nova] Lock "0eb943ae-462c-40dd-a870-7a884ea2db74-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.003s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.633970] env[62965]: DEBUG oslo_concurrency.lockutils [req-4ca6a2d1-753f-426a-b542-cfc0f777c991 req-6f38726d-1b11-4012-844f-38c0fb6ecd7b service nova] Lock "0eb943ae-462c-40dd-a870-7a884ea2db74-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.634195] env[62965]: DEBUG nova.compute.manager [req-4ca6a2d1-753f-426a-b542-cfc0f777c991 req-6f38726d-1b11-4012-844f-38c0fb6ecd7b service nova] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] No waiting events found dispatching network-vif-plugged-c6f37c93-3ad8-4408-9f81-519246ae7051 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 924.634372] env[62965]: WARNING nova.compute.manager [req-4ca6a2d1-753f-426a-b542-cfc0f777c991 req-6f38726d-1b11-4012-844f-38c0fb6ecd7b service nova] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Received unexpected event network-vif-plugged-c6f37c93-3ad8-4408-9f81-519246ae7051 for instance with vm_state building and task_state spawning. [ 924.723275] env[62965]: DEBUG nova.compute.utils [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 924.727271] env[62965]: INFO nova.compute.claims [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 924.731157] env[62965]: DEBUG nova.compute.manager [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 924.731343] env[62965]: DEBUG nova.network.neutron [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 924.800660] env[62965]: DEBUG nova.policy [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '97ee4fdb7cdb45829c7fe694e883da8c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1ed71548c085499981fa1b102d660368', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 924.825916] env[62965]: INFO nova.compute.manager [None req-ce981bd3-6c64-4571-885c-7e3dd74e2d52 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Detaching volume eeae4205-49ab-42e1-9d79-4b17b3886993 [ 924.869616] env[62965]: DEBUG oslo_vmware.api [None req-23c6bb8c-957a-4b64-9cee-aa35e0118c0a tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390044, 'name': PowerOffVM_Task, 'duration_secs': 0.394548} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.869790] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-23c6bb8c-957a-4b64-9cee-aa35e0118c0a tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 924.870123] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-23c6bb8c-957a-4b64-9cee-aa35e0118c0a tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 924.870304] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0e17fca7-f7a4-4f24-92d9-0675f217bff1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.872952] env[62965]: INFO nova.virt.block_device [None req-ce981bd3-6c64-4571-885c-7e3dd74e2d52 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Attempting to driver detach volume eeae4205-49ab-42e1-9d79-4b17b3886993 from mountpoint /dev/sdb [ 924.873185] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce981bd3-6c64-4571-885c-7e3dd74e2d52 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Volume detach. Driver type: vmdk {{(pid=62965) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 924.873373] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce981bd3-6c64-4571-885c-7e3dd74e2d52 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-295006', 'volume_id': 'eeae4205-49ab-42e1-9d79-4b17b3886993', 'name': 'volume-eeae4205-49ab-42e1-9d79-4b17b3886993', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '64cf898c-2c97-4daa-80c8-0a1c252cbcf2', 'attached_at': '', 'detached_at': '', 'volume_id': 'eeae4205-49ab-42e1-9d79-4b17b3886993', 'serial': 'eeae4205-49ab-42e1-9d79-4b17b3886993'} {{(pid=62965) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 924.874560] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8779249a-0e14-4ca8-8ce0-2ef006550028 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.906472] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-182a1786-2889-432e-b9a6-28874134b309 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.914961] env[62965]: DEBUG oslo_vmware.api [None req-36b4d40d-b72a-4a0e-a875-67e53da32f82 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390042, 'name': ReconfigVM_Task, 'duration_secs': 0.836098} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.916943] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-36b4d40d-b72a-4a0e-a875-67e53da32f82 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-295011', 'volume_id': 'edb73cdb-c23b-4dae-a3b5-8f4c025464d7', 'name': 'volume-edb73cdb-c23b-4dae-a3b5-8f4c025464d7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6def4a38-d3b8-4724-97a1-1859b239d900', 'attached_at': '', 'detached_at': '', 'volume_id': 'edb73cdb-c23b-4dae-a3b5-8f4c025464d7', 'serial': 'edb73cdb-c23b-4dae-a3b5-8f4c025464d7'} {{(pid=62965) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 924.919793] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a12ca57-0d3c-4594-be61-c8cf82f0dabf {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.950130] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa52ce8d-5384-4f0d-8f6f-42b959443255 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.952760] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-23c6bb8c-957a-4b64-9cee-aa35e0118c0a tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 924.952962] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-23c6bb8c-957a-4b64-9cee-aa35e0118c0a tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Deleting contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 924.953202] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-23c6bb8c-957a-4b64-9cee-aa35e0118c0a tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Deleting the datastore file [datastore2] e0c4e188-a041-4aa7-95de-6e8718b17ade {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 924.953603] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f02e662a-64f0-41ed-bd08-e354bf0e9c05 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.959755] env[62965]: DEBUG oslo_vmware.api [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390043, 'name': CloneVM_Task} progress is 94%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.971838] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce981bd3-6c64-4571-885c-7e3dd74e2d52 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] The volume has not been displaced from its original location: [datastore2] volume-eeae4205-49ab-42e1-9d79-4b17b3886993/volume-eeae4205-49ab-42e1-9d79-4b17b3886993.vmdk. No consolidation needed. {{(pid=62965) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 924.977129] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce981bd3-6c64-4571-885c-7e3dd74e2d52 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Reconfiguring VM instance instance-00000040 to detach disk 2001 {{(pid=62965) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 924.977525] env[62965]: DEBUG oslo_vmware.api [None req-23c6bb8c-957a-4b64-9cee-aa35e0118c0a tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 924.977525] env[62965]: value = "task-1390046" [ 924.977525] env[62965]: _type = "Task" [ 924.977525] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.979932] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0eb69e36-f26c-4892-a133-0897172bfe13 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.002688] env[62965]: DEBUG oslo_vmware.api [None req-23c6bb8c-957a-4b64-9cee-aa35e0118c0a tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390046, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.004160] env[62965]: DEBUG oslo_vmware.api [None req-ce981bd3-6c64-4571-885c-7e3dd74e2d52 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Waiting for the task: (returnval){ [ 925.004160] env[62965]: value = "task-1390047" [ 925.004160] env[62965]: _type = "Task" [ 925.004160] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.007552] env[62965]: DEBUG oslo_concurrency.lockutils [req-2c42f935-b0e3-4178-89da-beb8f378b5ff req-ebce7590-8c2f-4d34-a1b5-7daa8e05a90d service nova] Releasing lock "refresh_cache-00f24976-2f0e-4a2f-8f2e-012b0fce479f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.012760] env[62965]: DEBUG oslo_vmware.api [None req-ce981bd3-6c64-4571-885c-7e3dd74e2d52 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1390047, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.059307] env[62965]: DEBUG nova.network.neutron [req-578e0ad4-7ef2-4b33-90d1-71ae5e4e5c64 req-13f4f758-f244-484b-8e53-95186ce60a04 service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Updated VIF entry in instance network info cache for port b1718061-edf1-4077-b57f-0e54aea365a6. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 925.059722] env[62965]: DEBUG nova.network.neutron [req-578e0ad4-7ef2-4b33-90d1-71ae5e4e5c64 req-13f4f758-f244-484b-8e53-95186ce60a04 service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Updating instance_info_cache with network_info: [{"id": "b1718061-edf1-4077-b57f-0e54aea365a6", "address": "fa:16:3e:1f:27:16", "network": {"id": "a9699d8b-5e76-4666-a236-33ab8f22345e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1329540850-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1952dedf340f4b7ba0d57af6bba9a749", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a316376e-2ef0-4b1e-b40c-10321ebd7e1a", "external-id": "nsx-vlan-transportzone-942", "segmentation_id": 942, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1718061-ed", "ovs_interfaceid": "b1718061-edf1-4077-b57f-0e54aea365a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.165593] env[62965]: DEBUG nova.network.neutron [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Successfully created port: 849e89a5-b45f-482f-a43b-4fd17a659c17 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 925.170147] env[62965]: DEBUG nova.network.neutron [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Successfully updated port: c6f37c93-3ad8-4408-9f81-519246ae7051 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 925.200366] env[62965]: DEBUG nova.compute.manager [req-ac601a19-8c43-4306-97d7-53611fc06c07 req-bdec33e4-d1c0-4763-9cbf-3ab3799eb0ac service nova] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Received event network-changed-c6f37c93-3ad8-4408-9f81-519246ae7051 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 925.200789] env[62965]: DEBUG nova.compute.manager [req-ac601a19-8c43-4306-97d7-53611fc06c07 req-bdec33e4-d1c0-4763-9cbf-3ab3799eb0ac service nova] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Refreshing instance network info cache due to event network-changed-c6f37c93-3ad8-4408-9f81-519246ae7051. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 925.200852] env[62965]: DEBUG oslo_concurrency.lockutils [req-ac601a19-8c43-4306-97d7-53611fc06c07 req-bdec33e4-d1c0-4763-9cbf-3ab3799eb0ac service nova] Acquiring lock "refresh_cache-0eb943ae-462c-40dd-a870-7a884ea2db74" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.200999] env[62965]: DEBUG oslo_concurrency.lockutils [req-ac601a19-8c43-4306-97d7-53611fc06c07 req-bdec33e4-d1c0-4763-9cbf-3ab3799eb0ac service nova] Acquired lock "refresh_cache-0eb943ae-462c-40dd-a870-7a884ea2db74" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.201233] env[62965]: DEBUG nova.network.neutron [req-ac601a19-8c43-4306-97d7-53611fc06c07 req-bdec33e4-d1c0-4763-9cbf-3ab3799eb0ac service nova] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Refreshing network info cache for port c6f37c93-3ad8-4408-9f81-519246ae7051 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 925.232354] env[62965]: DEBUG nova.compute.manager [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 925.236637] env[62965]: INFO nova.compute.resource_tracker [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Updating resource usage from migration 5db4bd92-be48-40f0-88ac-7cda961107d8 [ 925.456722] env[62965]: DEBUG oslo_vmware.api [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390043, 'name': CloneVM_Task, 'duration_secs': 1.454155} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.462570] env[62965]: INFO nova.virt.vmwareapi.vmops [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Created linked-clone VM from snapshot [ 925.462570] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-077fbd8e-10c6-4dce-a0b1-5d8b40f10f4e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.464541] env[62965]: DEBUG nova.objects.instance [None req-36b4d40d-b72a-4a0e-a875-67e53da32f82 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lazy-loading 'flavor' on Instance uuid 6def4a38-d3b8-4724-97a1-1859b239d900 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 925.471355] env[62965]: DEBUG nova.virt.vmwareapi.images [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Uploading image c93dd56b-480e-46d6-9aed-4d5aa28dd1e0 {{(pid=62965) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 925.475877] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec2a4c97-2aa0-4328-8aee-d09ec7977c49 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.484318] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f014e8b7-31d0-4fc6-8c68-308c82fefce5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.515807] env[62965]: DEBUG oslo_vmware.rw_handles [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 925.515807] env[62965]: value = "vm-295016" [ 925.515807] env[62965]: _type = "VirtualMachine" [ 925.515807] env[62965]: }. {{(pid=62965) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 925.521860] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3662f22c-94e7-4b62-b4bb-8584adcd1db9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.523964] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-dff595c9-17c8-4381-9237-e37edc3831f4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.536138] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab4cef45-51d2-4828-9123-c2aee2c43717 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.539786] env[62965]: DEBUG oslo_vmware.api [None req-23c6bb8c-957a-4b64-9cee-aa35e0118c0a tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390046, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.194618} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.540055] env[62965]: DEBUG oslo_vmware.api [None req-ce981bd3-6c64-4571-885c-7e3dd74e2d52 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1390047, 'name': ReconfigVM_Task, 'duration_secs': 0.266359} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.541759] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-23c6bb8c-957a-4b64-9cee-aa35e0118c0a tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 925.541953] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-23c6bb8c-957a-4b64-9cee-aa35e0118c0a tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Deleted contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 925.542154] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-23c6bb8c-957a-4b64-9cee-aa35e0118c0a tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 925.542331] env[62965]: INFO nova.compute.manager [None req-23c6bb8c-957a-4b64-9cee-aa35e0118c0a tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Took 1.20 seconds to destroy the instance on the hypervisor. [ 925.542568] env[62965]: DEBUG oslo.service.loopingcall [None req-23c6bb8c-957a-4b64-9cee-aa35e0118c0a tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 925.542774] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce981bd3-6c64-4571-885c-7e3dd74e2d52 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Reconfigured VM instance instance-00000040 to detach disk 2001 {{(pid=62965) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 925.548324] env[62965]: DEBUG nova.compute.manager [-] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 925.548429] env[62965]: DEBUG nova.network.neutron [-] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 925.550189] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-29689aae-22b0-4e8e-b401-a9e5e008d682 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.568914] env[62965]: DEBUG oslo_vmware.rw_handles [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lease: (returnval){ [ 925.568914] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52e1929f-4797-c81c-c0e1-42b2959a9069" [ 925.568914] env[62965]: _type = "HttpNfcLease" [ 925.568914] env[62965]: } obtained for exporting VM: (result){ [ 925.568914] env[62965]: value = "vm-295016" [ 925.568914] env[62965]: _type = "VirtualMachine" [ 925.568914] env[62965]: }. {{(pid=62965) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 925.569168] env[62965]: DEBUG oslo_vmware.api [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the lease: (returnval){ [ 925.569168] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52e1929f-4797-c81c-c0e1-42b2959a9069" [ 925.569168] env[62965]: _type = "HttpNfcLease" [ 925.569168] env[62965]: } to be ready. {{(pid=62965) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 925.569582] env[62965]: DEBUG oslo_concurrency.lockutils [req-578e0ad4-7ef2-4b33-90d1-71ae5e4e5c64 req-13f4f758-f244-484b-8e53-95186ce60a04 service nova] Releasing lock "refresh_cache-d038608c-9264-4fd5-b9ba-0653c98ec0e4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.570200] env[62965]: DEBUG nova.compute.provider_tree [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 925.577285] env[62965]: DEBUG oslo_vmware.api [None req-ce981bd3-6c64-4571-885c-7e3dd74e2d52 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Waiting for the task: (returnval){ [ 925.577285] env[62965]: value = "task-1390049" [ 925.577285] env[62965]: _type = "Task" [ 925.577285] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.579013] env[62965]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 925.579013] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52e1929f-4797-c81c-c0e1-42b2959a9069" [ 925.579013] env[62965]: _type = "HttpNfcLease" [ 925.579013] env[62965]: } is ready. {{(pid=62965) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 925.582134] env[62965]: DEBUG oslo_vmware.rw_handles [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 925.582134] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52e1929f-4797-c81c-c0e1-42b2959a9069" [ 925.582134] env[62965]: _type = "HttpNfcLease" [ 925.582134] env[62965]: }. {{(pid=62965) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 925.583350] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb94e561-2185-4f43-b5ec-429b1432a1e1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.593029] env[62965]: DEBUG oslo_vmware.rw_handles [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528b67d5-b4d9-9eb3-374d-7be210b76da1/disk-0.vmdk from lease info. {{(pid=62965) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 925.593224] env[62965]: DEBUG oslo_vmware.rw_handles [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528b67d5-b4d9-9eb3-374d-7be210b76da1/disk-0.vmdk for reading. {{(pid=62965) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 925.599823] env[62965]: DEBUG oslo_vmware.api [None req-ce981bd3-6c64-4571-885c-7e3dd74e2d52 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1390049, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.674693] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "refresh_cache-0eb943ae-462c-40dd-a870-7a884ea2db74" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.699073] env[62965]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-6f1bebb7-b3f5-4bdd-8dc3-e24acb78c0ab {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.747037] env[62965]: DEBUG nova.network.neutron [req-ac601a19-8c43-4306-97d7-53611fc06c07 req-bdec33e4-d1c0-4763-9cbf-3ab3799eb0ac service nova] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 925.879040] env[62965]: DEBUG nova.network.neutron [req-ac601a19-8c43-4306-97d7-53611fc06c07 req-bdec33e4-d1c0-4763-9cbf-3ab3799eb0ac service nova] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.073741] env[62965]: DEBUG nova.scheduler.client.report [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 926.089559] env[62965]: DEBUG oslo_vmware.api [None req-ce981bd3-6c64-4571-885c-7e3dd74e2d52 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1390049, 'name': ReconfigVM_Task, 'duration_secs': 0.232861} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.090037] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce981bd3-6c64-4571-885c-7e3dd74e2d52 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-295006', 'volume_id': 'eeae4205-49ab-42e1-9d79-4b17b3886993', 'name': 'volume-eeae4205-49ab-42e1-9d79-4b17b3886993', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '64cf898c-2c97-4daa-80c8-0a1c252cbcf2', 'attached_at': '', 'detached_at': '', 'volume_id': 'eeae4205-49ab-42e1-9d79-4b17b3886993', 'serial': 'eeae4205-49ab-42e1-9d79-4b17b3886993'} {{(pid=62965) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 926.251458] env[62965]: DEBUG nova.compute.manager [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 926.275175] env[62965]: DEBUG nova.virt.hardware [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 926.276078] env[62965]: DEBUG nova.virt.hardware [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 926.276078] env[62965]: DEBUG nova.virt.hardware [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 926.277020] env[62965]: DEBUG nova.virt.hardware [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 926.277020] env[62965]: DEBUG nova.virt.hardware [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 926.277020] env[62965]: DEBUG nova.virt.hardware [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 926.277020] env[62965]: DEBUG nova.virt.hardware [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 926.277202] env[62965]: DEBUG nova.virt.hardware [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 926.277367] env[62965]: DEBUG nova.virt.hardware [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 926.277536] env[62965]: DEBUG nova.virt.hardware [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 926.277715] env[62965]: DEBUG nova.virt.hardware [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 926.278616] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-609fa844-e597-464b-afbc-a0de2278ceb2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.286554] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dd56e51-da37-43bd-8065-0d3a53a8735e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.382872] env[62965]: DEBUG oslo_concurrency.lockutils [req-ac601a19-8c43-4306-97d7-53611fc06c07 req-bdec33e4-d1c0-4763-9cbf-3ab3799eb0ac service nova] Releasing lock "refresh_cache-0eb943ae-462c-40dd-a870-7a884ea2db74" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.382872] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquired lock "refresh_cache-0eb943ae-462c-40dd-a870-7a884ea2db74" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.382872] env[62965]: DEBUG nova.network.neutron [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 926.436485] env[62965]: DEBUG nova.network.neutron [-] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.472129] env[62965]: DEBUG oslo_concurrency.lockutils [None req-36b4d40d-b72a-4a0e-a875-67e53da32f82 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "6def4a38-d3b8-4724-97a1-1859b239d900" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.328s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.579049] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.359s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.579306] env[62965]: INFO nova.compute.manager [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Migrating [ 926.579552] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.579727] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquired lock "compute-rpcapi-router" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.581272] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4a798d04-cb0c-4ce1-925f-11f6078cb448 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.155s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.581578] env[62965]: DEBUG nova.objects.instance [None req-4a798d04-cb0c-4ce1-925f-11f6078cb448 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lazy-loading 'resources' on Instance uuid f82d86ec-3364-449d-be54-35627082c02b {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 926.644729] env[62965]: DEBUG nova.objects.instance [None req-ce981bd3-6c64-4571-885c-7e3dd74e2d52 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Lazy-loading 'flavor' on Instance uuid 64cf898c-2c97-4daa-80c8-0a1c252cbcf2 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 926.716317] env[62965]: DEBUG nova.network.neutron [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Successfully updated port: 849e89a5-b45f-482f-a43b-4fd17a659c17 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 926.914374] env[62965]: DEBUG nova.network.neutron [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 926.939997] env[62965]: INFO nova.compute.manager [-] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Took 1.39 seconds to deallocate network for instance. [ 927.066312] env[62965]: DEBUG nova.network.neutron [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Updating instance_info_cache with network_info: [{"id": "c6f37c93-3ad8-4408-9f81-519246ae7051", "address": "fa:16:3e:06:89:bb", "network": {"id": "62052700-1781-4933-93c8-fa1aafdb0fd2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1931798228-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "978666785b0743bb843de96585f96fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6f37c93-3a", "ovs_interfaceid": "c6f37c93-3ad8-4408-9f81-519246ae7051", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.083804] env[62965]: INFO nova.compute.rpcapi [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Automatically selected compute RPC version 6.4 from minimum service version 68 [ 927.084536] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Releasing lock "compute-rpcapi-router" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.219733] env[62965]: DEBUG oslo_concurrency.lockutils [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "refresh_cache-1be808db-d619-4c56-8092-ca52809ec10e" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.219733] env[62965]: DEBUG oslo_concurrency.lockutils [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquired lock "refresh_cache-1be808db-d619-4c56-8092-ca52809ec10e" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.220238] env[62965]: DEBUG nova.network.neutron [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 927.358248] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ccce397-f0fb-438a-8767-ad742886392e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.367230] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2ec5f0b-4550-452a-a86c-00b2b6fe655c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.400694] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1af15ff0-1671-4e66-a05a-f102bc8b157f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.409767] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9823dfa0-b57b-436e-93ab-0ff2b3ba49ec {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.424995] env[62965]: DEBUG nova.compute.provider_tree [None req-4a798d04-cb0c-4ce1-925f-11f6078cb448 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 927.446798] env[62965]: DEBUG oslo_concurrency.lockutils [None req-23c6bb8c-957a-4b64-9cee-aa35e0118c0a tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.464683] env[62965]: DEBUG oslo_concurrency.lockutils [None req-718b7f9d-fee5-4368-8c0b-8ca81b717274 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquiring lock "6def4a38-d3b8-4724-97a1-1859b239d900" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.464683] env[62965]: DEBUG oslo_concurrency.lockutils [None req-718b7f9d-fee5-4368-8c0b-8ca81b717274 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "6def4a38-d3b8-4724-97a1-1859b239d900" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.464683] env[62965]: DEBUG oslo_concurrency.lockutils [None req-718b7f9d-fee5-4368-8c0b-8ca81b717274 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquiring lock "6def4a38-d3b8-4724-97a1-1859b239d900-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.464683] env[62965]: DEBUG oslo_concurrency.lockutils [None req-718b7f9d-fee5-4368-8c0b-8ca81b717274 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "6def4a38-d3b8-4724-97a1-1859b239d900-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.464867] env[62965]: DEBUG oslo_concurrency.lockutils [None req-718b7f9d-fee5-4368-8c0b-8ca81b717274 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "6def4a38-d3b8-4724-97a1-1859b239d900-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.471039] env[62965]: INFO nova.compute.manager [None req-718b7f9d-fee5-4368-8c0b-8ca81b717274 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Terminating instance [ 927.568897] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Releasing lock "refresh_cache-0eb943ae-462c-40dd-a870-7a884ea2db74" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.569261] env[62965]: DEBUG nova.compute.manager [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Instance network_info: |[{"id": "c6f37c93-3ad8-4408-9f81-519246ae7051", "address": "fa:16:3e:06:89:bb", "network": {"id": "62052700-1781-4933-93c8-fa1aafdb0fd2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1931798228-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "978666785b0743bb843de96585f96fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc6f37c93-3a", "ovs_interfaceid": "c6f37c93-3ad8-4408-9f81-519246ae7051", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 927.569842] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:06:89:bb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a10c88d7-d13f-44fd-acee-7a734eb5f56a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c6f37c93-3ad8-4408-9f81-519246ae7051', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 927.577980] env[62965]: DEBUG oslo.service.loopingcall [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 927.578304] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 927.578607] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-89b3c858-08bb-4668-88ee-0133f2cc624a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.595922] env[62965]: DEBUG nova.compute.manager [req-f28a2710-9b74-43db-be1a-2317dbd0574c req-f2af5a10-e34d-4e45-882e-6fa762cdb627 service nova] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Received event network-vif-deleted-fe1bf4cc-7773-462b-b8b5-1849c661d910 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 927.596197] env[62965]: DEBUG nova.compute.manager [req-f28a2710-9b74-43db-be1a-2317dbd0574c req-f2af5a10-e34d-4e45-882e-6fa762cdb627 service nova] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Received event network-vif-plugged-849e89a5-b45f-482f-a43b-4fd17a659c17 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 927.596415] env[62965]: DEBUG oslo_concurrency.lockutils [req-f28a2710-9b74-43db-be1a-2317dbd0574c req-f2af5a10-e34d-4e45-882e-6fa762cdb627 service nova] Acquiring lock "1be808db-d619-4c56-8092-ca52809ec10e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.596689] env[62965]: DEBUG oslo_concurrency.lockutils [req-f28a2710-9b74-43db-be1a-2317dbd0574c req-f2af5a10-e34d-4e45-882e-6fa762cdb627 service nova] Lock "1be808db-d619-4c56-8092-ca52809ec10e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.596925] env[62965]: DEBUG oslo_concurrency.lockutils [req-f28a2710-9b74-43db-be1a-2317dbd0574c req-f2af5a10-e34d-4e45-882e-6fa762cdb627 service nova] Lock "1be808db-d619-4c56-8092-ca52809ec10e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.597149] env[62965]: DEBUG nova.compute.manager [req-f28a2710-9b74-43db-be1a-2317dbd0574c req-f2af5a10-e34d-4e45-882e-6fa762cdb627 service nova] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] No waiting events found dispatching network-vif-plugged-849e89a5-b45f-482f-a43b-4fd17a659c17 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 927.597374] env[62965]: WARNING nova.compute.manager [req-f28a2710-9b74-43db-be1a-2317dbd0574c req-f2af5a10-e34d-4e45-882e-6fa762cdb627 service nova] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Received unexpected event network-vif-plugged-849e89a5-b45f-482f-a43b-4fd17a659c17 for instance with vm_state building and task_state spawning. [ 927.597571] env[62965]: DEBUG nova.compute.manager [req-f28a2710-9b74-43db-be1a-2317dbd0574c req-f2af5a10-e34d-4e45-882e-6fa762cdb627 service nova] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Received event network-changed-849e89a5-b45f-482f-a43b-4fd17a659c17 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 927.597792] env[62965]: DEBUG nova.compute.manager [req-f28a2710-9b74-43db-be1a-2317dbd0574c req-f2af5a10-e34d-4e45-882e-6fa762cdb627 service nova] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Refreshing instance network info cache due to event network-changed-849e89a5-b45f-482f-a43b-4fd17a659c17. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 927.598035] env[62965]: DEBUG oslo_concurrency.lockutils [req-f28a2710-9b74-43db-be1a-2317dbd0574c req-f2af5a10-e34d-4e45-882e-6fa762cdb627 service nova] Acquiring lock "refresh_cache-1be808db-d619-4c56-8092-ca52809ec10e" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.604690] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "refresh_cache-fbbcbe60-92f3-497d-b190-520a2547b3a9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.604852] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquired lock "refresh_cache-fbbcbe60-92f3-497d-b190-520a2547b3a9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.605030] env[62965]: DEBUG nova.network.neutron [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 927.606209] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 927.606209] env[62965]: value = "task-1390050" [ 927.606209] env[62965]: _type = "Task" [ 927.606209] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.616029] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390050, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.654701] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ce981bd3-6c64-4571-885c-7e3dd74e2d52 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Lock "64cf898c-2c97-4daa-80c8-0a1c252cbcf2" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.334s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.714097] env[62965]: DEBUG oslo_concurrency.lockutils [None req-6a0685fb-b548-4893-a5d8-c149cc8c3463 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Acquiring lock "64cf898c-2c97-4daa-80c8-0a1c252cbcf2" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.714446] env[62965]: DEBUG oslo_concurrency.lockutils [None req-6a0685fb-b548-4893-a5d8-c149cc8c3463 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Lock "64cf898c-2c97-4daa-80c8-0a1c252cbcf2" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.759267] env[62965]: DEBUG nova.network.neutron [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 927.914411] env[62965]: DEBUG nova.network.neutron [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Updating instance_info_cache with network_info: [{"id": "849e89a5-b45f-482f-a43b-4fd17a659c17", "address": "fa:16:3e:f7:8d:59", "network": {"id": "a0a5e605-d580-4b7b-b3c0-5c7395bd5a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1679618017-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ed71548c085499981fa1b102d660368", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap849e89a5-b4", "ovs_interfaceid": "849e89a5-b45f-482f-a43b-4fd17a659c17", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.928336] env[62965]: DEBUG nova.scheduler.client.report [None req-4a798d04-cb0c-4ce1-925f-11f6078cb448 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 927.971270] env[62965]: DEBUG nova.compute.manager [None req-718b7f9d-fee5-4368-8c0b-8ca81b717274 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 927.971512] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-718b7f9d-fee5-4368-8c0b-8ca81b717274 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 927.972559] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5795704a-6130-4864-b245-ef3ce9b9dc0a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.980500] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-718b7f9d-fee5-4368-8c0b-8ca81b717274 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 927.980815] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-68a7f957-10bb-4209-915a-98b7ba1a1e62 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.986638] env[62965]: DEBUG oslo_vmware.api [None req-718b7f9d-fee5-4368-8c0b-8ca81b717274 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Waiting for the task: (returnval){ [ 927.986638] env[62965]: value = "task-1390051" [ 927.986638] env[62965]: _type = "Task" [ 927.986638] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.994664] env[62965]: DEBUG oslo_vmware.api [None req-718b7f9d-fee5-4368-8c0b-8ca81b717274 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390051, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.121060] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390050, 'name': CreateVM_Task, 'duration_secs': 0.352049} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.121060] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 928.121806] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.122269] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.123155] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 928.123609] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9885296f-8986-4c3a-8769-065db486db90 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.130053] env[62965]: DEBUG oslo_vmware.api [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 928.130053] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]529ed6c5-7e96-8aac-8b75-0d254bb7f835" [ 928.130053] env[62965]: _type = "Task" [ 928.130053] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.138609] env[62965]: DEBUG oslo_vmware.api [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]529ed6c5-7e96-8aac-8b75-0d254bb7f835, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.218038] env[62965]: INFO nova.compute.manager [None req-6a0685fb-b548-4893-a5d8-c149cc8c3463 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Detaching volume c5b95ab6-e0bb-4fcd-a82d-6b0a64749f14 [ 928.266054] env[62965]: INFO nova.virt.block_device [None req-6a0685fb-b548-4893-a5d8-c149cc8c3463 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Attempting to driver detach volume c5b95ab6-e0bb-4fcd-a82d-6b0a64749f14 from mountpoint /dev/sdc [ 928.267194] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a0685fb-b548-4893-a5d8-c149cc8c3463 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Volume detach. Driver type: vmdk {{(pid=62965) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 928.267194] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a0685fb-b548-4893-a5d8-c149cc8c3463 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-295012', 'volume_id': 'c5b95ab6-e0bb-4fcd-a82d-6b0a64749f14', 'name': 'volume-c5b95ab6-e0bb-4fcd-a82d-6b0a64749f14', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '64cf898c-2c97-4daa-80c8-0a1c252cbcf2', 'attached_at': '', 'detached_at': '', 'volume_id': 'c5b95ab6-e0bb-4fcd-a82d-6b0a64749f14', 'serial': 'c5b95ab6-e0bb-4fcd-a82d-6b0a64749f14'} {{(pid=62965) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 928.269026] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-102590f5-5fe6-4b10-85f7-dfa25a0ab546 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.292770] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce083fc6-09cd-48cd-8ed5-974b9c3b04f9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.301628] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc06111b-9971-4493-b309-b0beb90412ab {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.327289] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e1a8979-3857-4e9b-9338-29f3aecf39db {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.344970] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a0685fb-b548-4893-a5d8-c149cc8c3463 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] The volume has not been displaced from its original location: [datastore2] volume-c5b95ab6-e0bb-4fcd-a82d-6b0a64749f14/volume-c5b95ab6-e0bb-4fcd-a82d-6b0a64749f14.vmdk. No consolidation needed. {{(pid=62965) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 928.350729] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a0685fb-b548-4893-a5d8-c149cc8c3463 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Reconfiguring VM instance instance-00000040 to detach disk 2002 {{(pid=62965) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 928.352291] env[62965]: DEBUG nova.network.neutron [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Updating instance_info_cache with network_info: [{"id": "3edb59ed-a2b8-4d31-b975-9c9d64c3a22b", "address": "fa:16:3e:92:26:4b", "network": {"id": "a5246cc1-6001-4c20-9724-fd36a0437582", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-644702631-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bfa5cfdc7d5496482d87675f2a122dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3edb59ed-a2", "ovs_interfaceid": "3edb59ed-a2b8-4d31-b975-9c9d64c3a22b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.353806] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c6f34c15-71bf-441d-87e2-88b9265a34c0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.368139] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Releasing lock "refresh_cache-fbbcbe60-92f3-497d-b190-520a2547b3a9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.376311] env[62965]: DEBUG oslo_vmware.api [None req-6a0685fb-b548-4893-a5d8-c149cc8c3463 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Waiting for the task: (returnval){ [ 928.376311] env[62965]: value = "task-1390052" [ 928.376311] env[62965]: _type = "Task" [ 928.376311] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.385854] env[62965]: DEBUG oslo_vmware.api [None req-6a0685fb-b548-4893-a5d8-c149cc8c3463 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1390052, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.417894] env[62965]: DEBUG oslo_concurrency.lockutils [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Releasing lock "refresh_cache-1be808db-d619-4c56-8092-ca52809ec10e" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.418494] env[62965]: DEBUG nova.compute.manager [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Instance network_info: |[{"id": "849e89a5-b45f-482f-a43b-4fd17a659c17", "address": "fa:16:3e:f7:8d:59", "network": {"id": "a0a5e605-d580-4b7b-b3c0-5c7395bd5a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1679618017-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ed71548c085499981fa1b102d660368", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap849e89a5-b4", "ovs_interfaceid": "849e89a5-b45f-482f-a43b-4fd17a659c17", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 928.418995] env[62965]: DEBUG oslo_concurrency.lockutils [req-f28a2710-9b74-43db-be1a-2317dbd0574c req-f2af5a10-e34d-4e45-882e-6fa762cdb627 service nova] Acquired lock "refresh_cache-1be808db-d619-4c56-8092-ca52809ec10e" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.419357] env[62965]: DEBUG nova.network.neutron [req-f28a2710-9b74-43db-be1a-2317dbd0574c req-f2af5a10-e34d-4e45-882e-6fa762cdb627 service nova] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Refreshing network info cache for port 849e89a5-b45f-482f-a43b-4fd17a659c17 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 928.421055] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f7:8d:59', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0cd5d325-3053-407e-a4ee-f627e82a23f9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '849e89a5-b45f-482f-a43b-4fd17a659c17', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 928.435208] env[62965]: DEBUG oslo.service.loopingcall [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 928.437500] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4a798d04-cb0c-4ce1-925f-11f6078cb448 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.856s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.440570] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 928.441330] env[62965]: DEBUG oslo_concurrency.lockutils [None req-23c6bb8c-957a-4b64-9cee-aa35e0118c0a tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.995s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.441957] env[62965]: DEBUG nova.objects.instance [None req-23c6bb8c-957a-4b64-9cee-aa35e0118c0a tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Lazy-loading 'resources' on Instance uuid e0c4e188-a041-4aa7-95de-6e8718b17ade {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 928.444658] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-82825978-a5a6-45f5-ae67-bfb1188196ed {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.470463] env[62965]: INFO nova.scheduler.client.report [None req-4a798d04-cb0c-4ce1-925f-11f6078cb448 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Deleted allocations for instance f82d86ec-3364-449d-be54-35627082c02b [ 928.481973] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 928.481973] env[62965]: value = "task-1390053" [ 928.481973] env[62965]: _type = "Task" [ 928.481973] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.493664] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390053, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.498981] env[62965]: DEBUG oslo_vmware.api [None req-718b7f9d-fee5-4368-8c0b-8ca81b717274 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390051, 'name': PowerOffVM_Task, 'duration_secs': 0.219822} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.499295] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-718b7f9d-fee5-4368-8c0b-8ca81b717274 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 928.499488] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-718b7f9d-fee5-4368-8c0b-8ca81b717274 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 928.499797] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a32fec73-1287-4de9-b7ff-0a7399316deb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.578216] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-718b7f9d-fee5-4368-8c0b-8ca81b717274 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 928.578525] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-718b7f9d-fee5-4368-8c0b-8ca81b717274 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Deleting contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 928.578631] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-718b7f9d-fee5-4368-8c0b-8ca81b717274 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Deleting the datastore file [datastore1] 6def4a38-d3b8-4724-97a1-1859b239d900 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 928.578932] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5ef7bb41-aa6b-41f4-af32-360719c9f491 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.586366] env[62965]: DEBUG oslo_vmware.api [None req-718b7f9d-fee5-4368-8c0b-8ca81b717274 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Waiting for the task: (returnval){ [ 928.586366] env[62965]: value = "task-1390055" [ 928.586366] env[62965]: _type = "Task" [ 928.586366] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.597348] env[62965]: DEBUG oslo_vmware.api [None req-718b7f9d-fee5-4368-8c0b-8ca81b717274 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390055, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.642257] env[62965]: DEBUG oslo_vmware.api [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]529ed6c5-7e96-8aac-8b75-0d254bb7f835, 'name': SearchDatastore_Task, 'duration_secs': 0.015238} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.642637] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.642955] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 928.643343] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.643516] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.643762] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 928.644123] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f3956a45-029d-4c03-af23-65f71e6b7181 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.654080] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 928.654425] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 928.656080] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-48d96754-34cf-4895-bb8e-079499cfe65d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.664026] env[62965]: DEBUG oslo_vmware.api [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 928.664026] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52eb7443-e24d-ca29-197c-2900abac242e" [ 928.664026] env[62965]: _type = "Task" [ 928.664026] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.676141] env[62965]: DEBUG oslo_vmware.api [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52eb7443-e24d-ca29-197c-2900abac242e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.888190] env[62965]: DEBUG oslo_vmware.api [None req-6a0685fb-b548-4893-a5d8-c149cc8c3463 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1390052, 'name': ReconfigVM_Task, 'duration_secs': 0.258535} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.888536] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a0685fb-b548-4893-a5d8-c149cc8c3463 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Reconfigured VM instance instance-00000040 to detach disk 2002 {{(pid=62965) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 928.894275] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-14d767e3-85a7-4bd8-9790-b181e7025ded {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.908874] env[62965]: DEBUG oslo_vmware.api [None req-6a0685fb-b548-4893-a5d8-c149cc8c3463 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Waiting for the task: (returnval){ [ 928.908874] env[62965]: value = "task-1390056" [ 928.908874] env[62965]: _type = "Task" [ 928.908874] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.917608] env[62965]: DEBUG oslo_vmware.api [None req-6a0685fb-b548-4893-a5d8-c149cc8c3463 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1390056, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.980982] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4a798d04-cb0c-4ce1-925f-11f6078cb448 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "f82d86ec-3364-449d-be54-35627082c02b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.587s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.992042] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390053, 'name': CreateVM_Task, 'duration_secs': 0.398561} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.992470] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 928.996034] env[62965]: DEBUG oslo_concurrency.lockutils [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.996034] env[62965]: DEBUG oslo_concurrency.lockutils [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.996184] env[62965]: DEBUG oslo_concurrency.lockutils [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 928.997023] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0133089c-6d9e-447d-ab32-1e9bc5adc7ed {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.007349] env[62965]: DEBUG oslo_vmware.api [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 929.007349] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5272a700-9da6-b973-7cfd-56b8736da31b" [ 929.007349] env[62965]: _type = "Task" [ 929.007349] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.015155] env[62965]: DEBUG oslo_vmware.api [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5272a700-9da6-b973-7cfd-56b8736da31b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.099154] env[62965]: DEBUG oslo_vmware.api [None req-718b7f9d-fee5-4368-8c0b-8ca81b717274 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390055, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.310138} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.099905] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-718b7f9d-fee5-4368-8c0b-8ca81b717274 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 929.099905] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-718b7f9d-fee5-4368-8c0b-8ca81b717274 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Deleted contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 929.099905] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-718b7f9d-fee5-4368-8c0b-8ca81b717274 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 929.100178] env[62965]: INFO nova.compute.manager [None req-718b7f9d-fee5-4368-8c0b-8ca81b717274 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Took 1.13 seconds to destroy the instance on the hypervisor. [ 929.100349] env[62965]: DEBUG oslo.service.loopingcall [None req-718b7f9d-fee5-4368-8c0b-8ca81b717274 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 929.103268] env[62965]: DEBUG nova.compute.manager [-] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 929.103378] env[62965]: DEBUG nova.network.neutron [-] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 929.182580] env[62965]: DEBUG oslo_vmware.api [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52eb7443-e24d-ca29-197c-2900abac242e, 'name': SearchDatastore_Task, 'duration_secs': 0.012023} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.184814] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93b9fa39-25b4-4581-b574-acad108b7e48 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.187796] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1597766b-4bfe-426c-b2e0-b5fbaaaba61a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.195580] env[62965]: DEBUG oslo_vmware.api [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 929.195580] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]521baee2-b3f8-78c3-3731-6e8fcfc6305b" [ 929.195580] env[62965]: _type = "Task" [ 929.195580] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.197083] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-556ab462-9c59-4bc8-b85c-69d78bd9b0d9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.212368] env[62965]: DEBUG oslo_vmware.api [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]521baee2-b3f8-78c3-3731-6e8fcfc6305b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.238152] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25cb1cf1-d89d-48e6-afb3-f20effaa050d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.241599] env[62965]: DEBUG nova.network.neutron [req-f28a2710-9b74-43db-be1a-2317dbd0574c req-f2af5a10-e34d-4e45-882e-6fa762cdb627 service nova] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Updated VIF entry in instance network info cache for port 849e89a5-b45f-482f-a43b-4fd17a659c17. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 929.241944] env[62965]: DEBUG nova.network.neutron [req-f28a2710-9b74-43db-be1a-2317dbd0574c req-f2af5a10-e34d-4e45-882e-6fa762cdb627 service nova] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Updating instance_info_cache with network_info: [{"id": "849e89a5-b45f-482f-a43b-4fd17a659c17", "address": "fa:16:3e:f7:8d:59", "network": {"id": "a0a5e605-d580-4b7b-b3c0-5c7395bd5a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1679618017-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ed71548c085499981fa1b102d660368", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap849e89a5-b4", "ovs_interfaceid": "849e89a5-b45f-482f-a43b-4fd17a659c17", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.251153] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21d6192c-b0d0-4c62-98e4-e34786aa8f60 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.263499] env[62965]: DEBUG nova.compute.provider_tree [None req-23c6bb8c-957a-4b64-9cee-aa35e0118c0a tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 929.423024] env[62965]: DEBUG oslo_vmware.api [None req-6a0685fb-b548-4893-a5d8-c149cc8c3463 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1390056, 'name': ReconfigVM_Task, 'duration_secs': 0.193403} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.423024] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a0685fb-b548-4893-a5d8-c149cc8c3463 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-295012', 'volume_id': 'c5b95ab6-e0bb-4fcd-a82d-6b0a64749f14', 'name': 'volume-c5b95ab6-e0bb-4fcd-a82d-6b0a64749f14', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '64cf898c-2c97-4daa-80c8-0a1c252cbcf2', 'attached_at': '', 'detached_at': '', 'volume_id': 'c5b95ab6-e0bb-4fcd-a82d-6b0a64749f14', 'serial': 'c5b95ab6-e0bb-4fcd-a82d-6b0a64749f14'} {{(pid=62965) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 929.517014] env[62965]: DEBUG oslo_vmware.api [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5272a700-9da6-b973-7cfd-56b8736da31b, 'name': SearchDatastore_Task, 'duration_secs': 0.01249} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.517375] env[62965]: DEBUG oslo_concurrency.lockutils [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.517612] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 929.517849] env[62965]: DEBUG oslo_concurrency.lockutils [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.565247] env[62965]: DEBUG oslo_concurrency.lockutils [None req-056ceac3-13ae-4a53-9d5d-071903fb9717 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquiring lock "d4d9b202-3aa7-4b66-949d-4a2eb34b3adf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.565514] env[62965]: DEBUG oslo_concurrency.lockutils [None req-056ceac3-13ae-4a53-9d5d-071903fb9717 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "d4d9b202-3aa7-4b66-949d-4a2eb34b3adf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.566313] env[62965]: DEBUG oslo_concurrency.lockutils [None req-056ceac3-13ae-4a53-9d5d-071903fb9717 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquiring lock "d4d9b202-3aa7-4b66-949d-4a2eb34b3adf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.566313] env[62965]: DEBUG oslo_concurrency.lockutils [None req-056ceac3-13ae-4a53-9d5d-071903fb9717 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "d4d9b202-3aa7-4b66-949d-4a2eb34b3adf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.566313] env[62965]: DEBUG oslo_concurrency.lockutils [None req-056ceac3-13ae-4a53-9d5d-071903fb9717 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "d4d9b202-3aa7-4b66-949d-4a2eb34b3adf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.570067] env[62965]: INFO nova.compute.manager [None req-056ceac3-13ae-4a53-9d5d-071903fb9717 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Terminating instance [ 929.711491] env[62965]: DEBUG oslo_vmware.api [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]521baee2-b3f8-78c3-3731-6e8fcfc6305b, 'name': SearchDatastore_Task, 'duration_secs': 0.018926} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.712197] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.715028] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 0eb943ae-462c-40dd-a870-7a884ea2db74/0eb943ae-462c-40dd-a870-7a884ea2db74.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 929.715028] env[62965]: DEBUG oslo_concurrency.lockutils [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.715028] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 929.715028] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2b32c910-216a-4121-b9fe-fcb9804d51b8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.717301] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2dcfa22d-6c5b-4747-ab95-7f55127b45bb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.724585] env[62965]: DEBUG oslo_vmware.api [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 929.724585] env[62965]: value = "task-1390057" [ 929.724585] env[62965]: _type = "Task" [ 929.724585] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.732040] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 929.732252] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 929.735151] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-530235a0-6a96-4da9-be70-2330a0633a82 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.740491] env[62965]: DEBUG oslo_vmware.api [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390057, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.743404] env[62965]: DEBUG oslo_vmware.api [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 929.743404] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]526249b6-4e76-c3e5-c524-2d3fdc24802a" [ 929.743404] env[62965]: _type = "Task" [ 929.743404] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.743921] env[62965]: DEBUG oslo_concurrency.lockutils [req-f28a2710-9b74-43db-be1a-2317dbd0574c req-f2af5a10-e34d-4e45-882e-6fa762cdb627 service nova] Releasing lock "refresh_cache-1be808db-d619-4c56-8092-ca52809ec10e" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.751263] env[62965]: DEBUG oslo_vmware.api [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]526249b6-4e76-c3e5-c524-2d3fdc24802a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.767256] env[62965]: DEBUG nova.scheduler.client.report [None req-23c6bb8c-957a-4b64-9cee-aa35e0118c0a tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 929.887025] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37a6f243-86ed-40ae-98e1-a37b0c32f147 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.908878] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Updating instance 'fbbcbe60-92f3-497d-b190-520a2547b3a9' progress to 0 {{(pid=62965) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 929.978446] env[62965]: DEBUG nova.objects.instance [None req-6a0685fb-b548-4893-a5d8-c149cc8c3463 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Lazy-loading 'flavor' on Instance uuid 64cf898c-2c97-4daa-80c8-0a1c252cbcf2 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 929.982521] env[62965]: DEBUG nova.compute.manager [req-f551f6f8-e05f-422b-b258-3f7c3af37290 req-8b098fa2-dd98-4a03-a3c4-2c5392fe6719 service nova] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Received event network-vif-deleted-65a91c43-8ff0-4e66-8f18-2de456875968 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 929.982606] env[62965]: INFO nova.compute.manager [req-f551f6f8-e05f-422b-b258-3f7c3af37290 req-8b098fa2-dd98-4a03-a3c4-2c5392fe6719 service nova] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Neutron deleted interface 65a91c43-8ff0-4e66-8f18-2de456875968; detaching it from the instance and deleting it from the info cache [ 929.982802] env[62965]: DEBUG nova.network.neutron [req-f551f6f8-e05f-422b-b258-3f7c3af37290 req-8b098fa2-dd98-4a03-a3c4-2c5392fe6719 service nova] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.077031] env[62965]: DEBUG nova.compute.manager [None req-056ceac3-13ae-4a53-9d5d-071903fb9717 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 930.077031] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-056ceac3-13ae-4a53-9d5d-071903fb9717 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 930.077031] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48b2bb60-95e4-43be-9516-2cb231773727 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.091364] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-056ceac3-13ae-4a53-9d5d-071903fb9717 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 930.091364] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-803d24eb-f732-44a7-b626-ddade4af1a07 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.102981] env[62965]: DEBUG oslo_vmware.api [None req-056ceac3-13ae-4a53-9d5d-071903fb9717 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 930.102981] env[62965]: value = "task-1390058" [ 930.102981] env[62965]: _type = "Task" [ 930.102981] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.116200] env[62965]: DEBUG oslo_vmware.api [None req-056ceac3-13ae-4a53-9d5d-071903fb9717 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390058, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.194632] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquiring lock "97954de1-7b4b-4615-8d0f-0e8580a515ed" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.194632] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "97954de1-7b4b-4615-8d0f-0e8580a515ed" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.239332] env[62965]: DEBUG oslo_vmware.api [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390057, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.259063] env[62965]: DEBUG oslo_vmware.api [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]526249b6-4e76-c3e5-c524-2d3fdc24802a, 'name': SearchDatastore_Task, 'duration_secs': 0.026783} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.260028] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f61d8bf-f2db-4399-8a6e-20d97dfae96c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.267605] env[62965]: DEBUG oslo_vmware.api [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 930.267605] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]520f6cd1-f0ff-df85-fb43-f699571fcdb9" [ 930.267605] env[62965]: _type = "Task" [ 930.267605] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.272578] env[62965]: DEBUG oslo_concurrency.lockutils [None req-23c6bb8c-957a-4b64-9cee-aa35e0118c0a tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.831s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.282468] env[62965]: DEBUG oslo_vmware.api [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]520f6cd1-f0ff-df85-fb43-f699571fcdb9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.297478] env[62965]: INFO nova.scheduler.client.report [None req-23c6bb8c-957a-4b64-9cee-aa35e0118c0a tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Deleted allocations for instance e0c4e188-a041-4aa7-95de-6e8718b17ade [ 930.374096] env[62965]: DEBUG nova.network.neutron [-] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.416218] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 930.416443] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3a9cb5d6-54b7-4e8b-9738-9b885c4903bf {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.423777] env[62965]: DEBUG oslo_vmware.api [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 930.423777] env[62965]: value = "task-1390059" [ 930.423777] env[62965]: _type = "Task" [ 930.423777] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.432789] env[62965]: DEBUG oslo_vmware.api [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390059, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.488139] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cd3220f5-5fab-4f9f-af6b-31cc925eb2ec {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.499270] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d0298cd-54d4-429a-abe6-d28417664497 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.536433] env[62965]: DEBUG nova.compute.manager [req-f551f6f8-e05f-422b-b258-3f7c3af37290 req-8b098fa2-dd98-4a03-a3c4-2c5392fe6719 service nova] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Detach interface failed, port_id=65a91c43-8ff0-4e66-8f18-2de456875968, reason: Instance 6def4a38-d3b8-4724-97a1-1859b239d900 could not be found. {{(pid=62965) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11294}} [ 930.614096] env[62965]: DEBUG oslo_vmware.api [None req-056ceac3-13ae-4a53-9d5d-071903fb9717 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390058, 'name': PowerOffVM_Task, 'duration_secs': 0.297656} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.614432] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-056ceac3-13ae-4a53-9d5d-071903fb9717 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 930.614661] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-056ceac3-13ae-4a53-9d5d-071903fb9717 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 930.614984] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3be37b3d-2bb5-49dd-ae69-f0242ed80c22 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.679209] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-056ceac3-13ae-4a53-9d5d-071903fb9717 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 930.679370] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-056ceac3-13ae-4a53-9d5d-071903fb9717 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Deleting contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 930.679503] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-056ceac3-13ae-4a53-9d5d-071903fb9717 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Deleting the datastore file [datastore2] d4d9b202-3aa7-4b66-949d-4a2eb34b3adf {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 930.679783] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dcbe2dba-fa84-4d02-929f-092ca4a5c787 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.685685] env[62965]: DEBUG oslo_vmware.api [None req-056ceac3-13ae-4a53-9d5d-071903fb9717 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 930.685685] env[62965]: value = "task-1390061" [ 930.685685] env[62965]: _type = "Task" [ 930.685685] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.693478] env[62965]: DEBUG oslo_vmware.api [None req-056ceac3-13ae-4a53-9d5d-071903fb9717 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390061, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.699015] env[62965]: DEBUG nova.compute.manager [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 930.734676] env[62965]: DEBUG oslo_vmware.api [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390057, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.59773} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.734957] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 0eb943ae-462c-40dd-a870-7a884ea2db74/0eb943ae-462c-40dd-a870-7a884ea2db74.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 930.735197] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 930.735463] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1ea4033e-af49-44fd-90ae-c64f232ecf5d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.741720] env[62965]: DEBUG oslo_vmware.api [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 930.741720] env[62965]: value = "task-1390062" [ 930.741720] env[62965]: _type = "Task" [ 930.741720] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.749717] env[62965]: DEBUG oslo_vmware.api [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390062, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.777705] env[62965]: DEBUG oslo_vmware.api [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]520f6cd1-f0ff-df85-fb43-f699571fcdb9, 'name': SearchDatastore_Task, 'duration_secs': 0.039119} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.778013] env[62965]: DEBUG oslo_concurrency.lockutils [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.778281] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 1be808db-d619-4c56-8092-ca52809ec10e/1be808db-d619-4c56-8092-ca52809ec10e.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 930.778576] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-62c39042-850c-4a5d-b85b-8f4af836567b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.784721] env[62965]: DEBUG oslo_vmware.api [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 930.784721] env[62965]: value = "task-1390063" [ 930.784721] env[62965]: _type = "Task" [ 930.784721] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.793147] env[62965]: DEBUG oslo_vmware.api [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390063, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.807251] env[62965]: DEBUG oslo_concurrency.lockutils [None req-23c6bb8c-957a-4b64-9cee-aa35e0118c0a tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Lock "e0c4e188-a041-4aa7-95de-6e8718b17ade" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.971s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.877055] env[62965]: INFO nova.compute.manager [-] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Took 1.77 seconds to deallocate network for instance. [ 930.933070] env[62965]: DEBUG oslo_vmware.api [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390059, 'name': PowerOffVM_Task, 'duration_secs': 0.183565} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.933364] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 930.933550] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Updating instance 'fbbcbe60-92f3-497d-b190-520a2547b3a9' progress to 17 {{(pid=62965) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 930.991306] env[62965]: DEBUG oslo_concurrency.lockutils [None req-6a0685fb-b548-4893-a5d8-c149cc8c3463 tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Lock "64cf898c-2c97-4daa-80c8-0a1c252cbcf2" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.277s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.196385] env[62965]: DEBUG oslo_vmware.api [None req-056ceac3-13ae-4a53-9d5d-071903fb9717 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390061, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.187059} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.196764] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-056ceac3-13ae-4a53-9d5d-071903fb9717 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 931.197031] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-056ceac3-13ae-4a53-9d5d-071903fb9717 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Deleted contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 931.197258] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-056ceac3-13ae-4a53-9d5d-071903fb9717 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 931.197444] env[62965]: INFO nova.compute.manager [None req-056ceac3-13ae-4a53-9d5d-071903fb9717 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Took 1.12 seconds to destroy the instance on the hypervisor. [ 931.197775] env[62965]: DEBUG oslo.service.loopingcall [None req-056ceac3-13ae-4a53-9d5d-071903fb9717 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 931.198014] env[62965]: DEBUG nova.compute.manager [-] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 931.198122] env[62965]: DEBUG nova.network.neutron [-] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 931.222230] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.222521] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.224383] env[62965]: INFO nova.compute.claims [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 931.253399] env[62965]: DEBUG oslo_vmware.api [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390062, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082571} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.253642] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 931.254316] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28c58486-72c7-4d10-8703-650841ec9e16 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.277351] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] 0eb943ae-462c-40dd-a870-7a884ea2db74/0eb943ae-462c-40dd-a870-7a884ea2db74.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 931.279856] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9d7b5b8f-8a29-49f4-9f51-65bdc878d946 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.305620] env[62965]: DEBUG oslo_vmware.api [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390063, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.307235] env[62965]: DEBUG oslo_vmware.api [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 931.307235] env[62965]: value = "task-1390064" [ 931.307235] env[62965]: _type = "Task" [ 931.307235] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.317770] env[62965]: DEBUG oslo_vmware.api [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390064, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.385612] env[62965]: DEBUG oslo_concurrency.lockutils [None req-718b7f9d-fee5-4368-8c0b-8ca81b717274 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.439956] env[62965]: DEBUG nova.virt.hardware [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:24Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 931.440372] env[62965]: DEBUG nova.virt.hardware [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 931.440487] env[62965]: DEBUG nova.virt.hardware [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 931.440652] env[62965]: DEBUG nova.virt.hardware [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 931.440863] env[62965]: DEBUG nova.virt.hardware [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 931.441110] env[62965]: DEBUG nova.virt.hardware [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 931.441296] env[62965]: DEBUG nova.virt.hardware [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 931.441498] env[62965]: DEBUG nova.virt.hardware [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 931.441691] env[62965]: DEBUG nova.virt.hardware [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 931.441893] env[62965]: DEBUG nova.virt.hardware [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 931.442150] env[62965]: DEBUG nova.virt.hardware [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 931.447337] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ac6ae3c5-14f8-45c4-95e6-682ed012a6e0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.466593] env[62965]: DEBUG oslo_vmware.api [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 931.466593] env[62965]: value = "task-1390065" [ 931.466593] env[62965]: _type = "Task" [ 931.466593] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.475569] env[62965]: DEBUG oslo_vmware.api [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390065, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.806097] env[62965]: DEBUG oslo_vmware.api [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390063, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.570197} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.806385] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 1be808db-d619-4c56-8092-ca52809ec10e/1be808db-d619-4c56-8092-ca52809ec10e.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 931.806602] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 931.806861] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-050f48ca-6772-49bf-8da5-d2c65d4c5c61 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.818931] env[62965]: DEBUG oslo_vmware.api [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390064, 'name': ReconfigVM_Task, 'duration_secs': 0.316271} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.820905] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Reconfigured VM instance instance-00000054 to attach disk [datastore2] 0eb943ae-462c-40dd-a870-7a884ea2db74/0eb943ae-462c-40dd-a870-7a884ea2db74.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 931.821026] env[62965]: DEBUG oslo_vmware.api [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 931.821026] env[62965]: value = "task-1390066" [ 931.821026] env[62965]: _type = "Task" [ 931.821026] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.821271] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0c538e3c-5a50-4dfa-8d02-46f82ecf9ae2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.831021] env[62965]: DEBUG oslo_vmware.api [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390066, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.832309] env[62965]: DEBUG oslo_vmware.api [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 931.832309] env[62965]: value = "task-1390067" [ 931.832309] env[62965]: _type = "Task" [ 931.832309] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.840592] env[62965]: DEBUG oslo_vmware.api [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390067, 'name': Rename_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.976495] env[62965]: DEBUG oslo_vmware.api [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390065, 'name': ReconfigVM_Task, 'duration_secs': 0.228011} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.976841] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Updating instance 'fbbcbe60-92f3-497d-b190-520a2547b3a9' progress to 33 {{(pid=62965) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 932.016884] env[62965]: DEBUG nova.network.neutron [-] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.172729] env[62965]: DEBUG nova.compute.manager [req-5343d435-76a9-410d-b6f5-6ac4194e16c7 req-3c6289a7-118f-4907-b670-7b132bd2cb06 service nova] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Received event network-vif-deleted-dc435eb8-9509-4115-b425-bbd397d12290 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 932.182603] env[62965]: DEBUG oslo_concurrency.lockutils [None req-48f705ca-e256-4c4e-a632-d7db59b218b0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "interface-d038608c-9264-4fd5-b9ba-0653c98ec0e4-9c46a8bf-6b23-45d3-95a4-f0c751895c71" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.182863] env[62965]: DEBUG oslo_concurrency.lockutils [None req-48f705ca-e256-4c4e-a632-d7db59b218b0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "interface-d038608c-9264-4fd5-b9ba-0653c98ec0e4-9c46a8bf-6b23-45d3-95a4-f0c751895c71" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.183231] env[62965]: DEBUG nova.objects.instance [None req-48f705ca-e256-4c4e-a632-d7db59b218b0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lazy-loading 'flavor' on Instance uuid d038608c-9264-4fd5-b9ba-0653c98ec0e4 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 932.227582] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Acquiring lock "3223b1ed-abc1-439c-bcad-6317eb204e45" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.227830] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Lock "3223b1ed-abc1-439c-bcad-6317eb204e45" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.333793] env[62965]: DEBUG oslo_vmware.api [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390066, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081881} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.339454] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 932.345940] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41dafa65-1a62-4e89-88db-b22febaaaa85 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.371160] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] 1be808db-d619-4c56-8092-ca52809ec10e/1be808db-d619-4c56-8092-ca52809ec10e.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 932.378421] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2dfa6571-2a39-42ee-9f96-bb727ba11e6a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.392983] env[62965]: DEBUG oslo_vmware.api [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390067, 'name': Rename_Task} progress is 99%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.400927] env[62965]: DEBUG oslo_vmware.api [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 932.400927] env[62965]: value = "task-1390068" [ 932.400927] env[62965]: _type = "Task" [ 932.400927] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.412845] env[62965]: DEBUG oslo_vmware.api [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390068, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.485215] env[62965]: DEBUG nova.virt.hardware [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 932.485215] env[62965]: DEBUG nova.virt.hardware [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 932.485215] env[62965]: DEBUG nova.virt.hardware [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 932.485215] env[62965]: DEBUG nova.virt.hardware [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 932.485604] env[62965]: DEBUG nova.virt.hardware [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 932.485604] env[62965]: DEBUG nova.virt.hardware [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 932.485604] env[62965]: DEBUG nova.virt.hardware [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 932.485717] env[62965]: DEBUG nova.virt.hardware [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 932.485964] env[62965]: DEBUG nova.virt.hardware [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 932.486213] env[62965]: DEBUG nova.virt.hardware [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 932.486453] env[62965]: DEBUG nova.virt.hardware [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 932.492991] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Reconfiguring VM instance instance-00000052 to detach disk 2000 {{(pid=62965) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 932.495902] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb34a99b-e003-400a-8644-67187b5076ce {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.517150] env[62965]: DEBUG oslo_vmware.api [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 932.517150] env[62965]: value = "task-1390069" [ 932.517150] env[62965]: _type = "Task" [ 932.517150] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.520770] env[62965]: INFO nova.compute.manager [-] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Took 1.32 seconds to deallocate network for instance. [ 932.524476] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18f26147-2d08-4905-ae99-4cb47e4f8d22 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.533861] env[62965]: DEBUG oslo_vmware.api [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390069, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.539483] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2669906-806c-4747-a96b-1e11b6a63615 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.572866] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77d81ed4-2dff-45a7-ac39-754a5e4b25a2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.582289] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9f5af7e-4bf6-4db6-9b13-7c60338477f5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.597940] env[62965]: DEBUG nova.compute.provider_tree [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 932.730627] env[62965]: DEBUG nova.compute.manager [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 932.786187] env[62965]: DEBUG nova.objects.instance [None req-48f705ca-e256-4c4e-a632-d7db59b218b0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lazy-loading 'pci_requests' on Instance uuid d038608c-9264-4fd5-b9ba-0653c98ec0e4 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 932.817807] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Acquiring lock "64cf898c-2c97-4daa-80c8-0a1c252cbcf2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.818151] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Lock "64cf898c-2c97-4daa-80c8-0a1c252cbcf2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.818409] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Acquiring lock "64cf898c-2c97-4daa-80c8-0a1c252cbcf2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.818549] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Lock "64cf898c-2c97-4daa-80c8-0a1c252cbcf2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.818730] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Lock "64cf898c-2c97-4daa-80c8-0a1c252cbcf2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.820935] env[62965]: INFO nova.compute.manager [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Terminating instance [ 932.844608] env[62965]: DEBUG oslo_vmware.api [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390067, 'name': Rename_Task} progress is 99%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.912203] env[62965]: DEBUG oslo_vmware.api [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390068, 'name': ReconfigVM_Task, 'duration_secs': 0.333479} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.912463] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Reconfigured VM instance instance-00000055 to attach disk [datastore2] 1be808db-d619-4c56-8092-ca52809ec10e/1be808db-d619-4c56-8092-ca52809ec10e.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 932.913097] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e6986205-7bf3-46e0-b2c4-c16a55f6b144 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.920965] env[62965]: DEBUG oslo_vmware.api [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 932.920965] env[62965]: value = "task-1390070" [ 932.920965] env[62965]: _type = "Task" [ 932.920965] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.930565] env[62965]: DEBUG oslo_vmware.api [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390070, 'name': Rename_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.027916] env[62965]: DEBUG oslo_vmware.api [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390069, 'name': ReconfigVM_Task, 'duration_secs': 0.192286} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.028387] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Reconfigured VM instance instance-00000052 to detach disk 2000 {{(pid=62965) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 933.029120] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edc24eea-2dcf-4350-b202-c49fc9e64f7c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.035341] env[62965]: DEBUG oslo_concurrency.lockutils [None req-056ceac3-13ae-4a53-9d5d-071903fb9717 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.054463] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Reconfiguring VM instance instance-00000052 to attach disk [datastore2] fbbcbe60-92f3-497d-b190-520a2547b3a9/fbbcbe60-92f3-497d-b190-520a2547b3a9.vmdk or device None with type thin {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 933.054799] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-880dbbc8-a157-4c59-a798-3d966aae42b9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.073579] env[62965]: DEBUG oslo_vmware.api [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 933.073579] env[62965]: value = "task-1390071" [ 933.073579] env[62965]: _type = "Task" [ 933.073579] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.082708] env[62965]: DEBUG oslo_vmware.api [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390071, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.101450] env[62965]: DEBUG nova.scheduler.client.report [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 933.251389] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.289348] env[62965]: DEBUG nova.objects.base [None req-48f705ca-e256-4c4e-a632-d7db59b218b0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62965) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 933.289596] env[62965]: DEBUG nova.network.neutron [None req-48f705ca-e256-4c4e-a632-d7db59b218b0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 933.325397] env[62965]: DEBUG nova.compute.manager [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 933.325700] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 933.326742] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12244ce5-5585-4f98-9130-c26e99eef4d3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.336856] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 933.340560] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ba7be805-93b5-4180-a759-a6c401b82593 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.348733] env[62965]: DEBUG oslo_vmware.api [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390067, 'name': Rename_Task, 'duration_secs': 1.169898} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.350217] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 933.350551] env[62965]: DEBUG oslo_vmware.api [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Waiting for the task: (returnval){ [ 933.350551] env[62965]: value = "task-1390072" [ 933.350551] env[62965]: _type = "Task" [ 933.350551] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.350817] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4bbf2456-8f1f-49d8-b662-bac125179611 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.362133] env[62965]: DEBUG oslo_vmware.api [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1390072, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.365102] env[62965]: DEBUG nova.policy [None req-48f705ca-e256-4c4e-a632-d7db59b218b0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '182d38ca18c64bed8f3ecc3f95229756', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1952dedf340f4b7ba0d57af6bba9a749', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 933.367173] env[62965]: DEBUG oslo_vmware.api [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 933.367173] env[62965]: value = "task-1390073" [ 933.367173] env[62965]: _type = "Task" [ 933.367173] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.377749] env[62965]: DEBUG oslo_vmware.api [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390073, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.431908] env[62965]: DEBUG oslo_vmware.api [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390070, 'name': Rename_Task, 'duration_secs': 0.146684} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.432296] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 933.432576] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f0de711d-02dd-4932-a257-649827f296cd {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.440756] env[62965]: DEBUG oslo_vmware.api [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 933.440756] env[62965]: value = "task-1390074" [ 933.440756] env[62965]: _type = "Task" [ 933.440756] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.452365] env[62965]: DEBUG oslo_vmware.api [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390074, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.586445] env[62965]: DEBUG oslo_vmware.api [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390071, 'name': ReconfigVM_Task, 'duration_secs': 0.277001} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.586801] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Reconfigured VM instance instance-00000052 to attach disk [datastore2] fbbcbe60-92f3-497d-b190-520a2547b3a9/fbbcbe60-92f3-497d-b190-520a2547b3a9.vmdk or device None with type thin {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 933.587066] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Updating instance 'fbbcbe60-92f3-497d-b190-520a2547b3a9' progress to 50 {{(pid=62965) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 933.607651] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.385s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.608541] env[62965]: DEBUG nova.compute.manager [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 933.611577] env[62965]: DEBUG oslo_concurrency.lockutils [None req-718b7f9d-fee5-4368-8c0b-8ca81b717274 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.226s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.611907] env[62965]: DEBUG nova.objects.instance [None req-718b7f9d-fee5-4368-8c0b-8ca81b717274 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lazy-loading 'resources' on Instance uuid 6def4a38-d3b8-4724-97a1-1859b239d900 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 933.866969] env[62965]: DEBUG oslo_vmware.api [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1390072, 'name': PowerOffVM_Task, 'duration_secs': 0.231202} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.867297] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 933.867566] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 933.867680] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8ffaf602-9788-48c8-9f6e-8da81c2653b1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.878598] env[62965]: DEBUG oslo_vmware.api [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390073, 'name': PowerOnVM_Task, 'duration_secs': 0.505049} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.878889] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 933.879117] env[62965]: INFO nova.compute.manager [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Took 10.03 seconds to spawn the instance on the hypervisor. [ 933.879372] env[62965]: DEBUG nova.compute.manager [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 933.880302] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-290bf7e5-bdc3-4087-a1b6-35efd92504fa {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.951956] env[62965]: DEBUG oslo_vmware.api [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390074, 'name': PowerOnVM_Task, 'duration_secs': 0.504889} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.952307] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 933.952658] env[62965]: INFO nova.compute.manager [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Took 7.70 seconds to spawn the instance on the hypervisor. [ 933.952815] env[62965]: DEBUG nova.compute.manager [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 933.953796] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ebbb1e2-7409-49d3-85b5-17fa6e4894e6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.971027] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 933.971935] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Deleting contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 933.971935] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Deleting the datastore file [datastore1] 64cf898c-2c97-4daa-80c8-0a1c252cbcf2 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 933.971935] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-18017594-5ffb-48d6-96aa-67e336bf6f22 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.979603] env[62965]: DEBUG oslo_vmware.api [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Waiting for the task: (returnval){ [ 933.979603] env[62965]: value = "task-1390076" [ 933.979603] env[62965]: _type = "Task" [ 933.979603] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.989173] env[62965]: DEBUG oslo_vmware.api [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1390076, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.094916] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96ecd13a-b386-46b8-b7bd-715b3f0207d7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.116118] env[62965]: DEBUG nova.compute.utils [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 934.120599] env[62965]: DEBUG nova.compute.manager [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 934.120854] env[62965]: DEBUG nova.network.neutron [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 934.123116] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c56d37cc-fc49-4254-ad73-29ef3fce8612 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.143032] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Updating instance 'fbbcbe60-92f3-497d-b190-520a2547b3a9' progress to 67 {{(pid=62965) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 934.165413] env[62965]: DEBUG nova.policy [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ce0e7f3429454fd18d54cbb964dfb561', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e1c6e8403409438f9ff10235cd2533bc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 934.342010] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bc77d68-f8d8-4c61-9667-1cd53619a36a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.351048] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a03b37a-7aeb-46b8-9636-384a877259f7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.388537] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3af5269-140d-4573-94b5-338bec8efe3d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.408887] env[62965]: INFO nova.compute.manager [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Took 17.79 seconds to build instance. [ 934.410418] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-157693dd-085c-4468-9d36-2002b91b2c65 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.430262] env[62965]: DEBUG nova.compute.provider_tree [None req-718b7f9d-fee5-4368-8c0b-8ca81b717274 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 934.438406] env[62965]: DEBUG oslo_vmware.rw_handles [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528b67d5-b4d9-9eb3-374d-7be210b76da1/disk-0.vmdk. {{(pid=62965) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 934.440085] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a948ee8a-d9f6-40e5-81ec-fde56494b6fc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.449375] env[62965]: DEBUG oslo_vmware.rw_handles [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528b67d5-b4d9-9eb3-374d-7be210b76da1/disk-0.vmdk is in state: ready. {{(pid=62965) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 934.449375] env[62965]: ERROR oslo_vmware.rw_handles [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528b67d5-b4d9-9eb3-374d-7be210b76da1/disk-0.vmdk due to incomplete transfer. [ 934.449375] env[62965]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-49250a90-c476-445a-8404-6d1fa93a5781 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.456502] env[62965]: DEBUG oslo_vmware.rw_handles [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528b67d5-b4d9-9eb3-374d-7be210b76da1/disk-0.vmdk. {{(pid=62965) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 934.456747] env[62965]: DEBUG nova.virt.vmwareapi.images [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Uploaded image c93dd56b-480e-46d6-9aed-4d5aa28dd1e0 to the Glance image server {{(pid=62965) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 934.460901] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Destroying the VM {{(pid=62965) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 934.460901] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-a163815e-82de-45c9-8d04-df93b2591db6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.472194] env[62965]: DEBUG oslo_vmware.api [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 934.472194] env[62965]: value = "task-1390077" [ 934.472194] env[62965]: _type = "Task" [ 934.472194] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.474793] env[62965]: DEBUG nova.network.neutron [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Successfully created port: 212a2219-fc24-44bb-8340-6ad61740edf0 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 934.480856] env[62965]: INFO nova.compute.manager [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Took 14.55 seconds to build instance. [ 934.492832] env[62965]: DEBUG oslo_vmware.api [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390077, 'name': Destroy_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.496624] env[62965]: DEBUG oslo_vmware.api [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Task: {'id': task-1390076, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.306811} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.497502] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 934.497721] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Deleted contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 934.497916] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 934.498156] env[62965]: INFO nova.compute.manager [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Took 1.17 seconds to destroy the instance on the hypervisor. [ 934.498417] env[62965]: DEBUG oslo.service.loopingcall [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 934.498929] env[62965]: DEBUG nova.compute.manager [-] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 934.499197] env[62965]: DEBUG nova.network.neutron [-] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 934.621630] env[62965]: DEBUG nova.compute.manager [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 934.692715] env[62965]: DEBUG nova.network.neutron [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Port 3edb59ed-a2b8-4d31-b975-9c9d64c3a22b binding to destination host cpu-1 is already ACTIVE {{(pid=62965) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 934.915609] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1a6a4864-5687-4ef4-b2c5-335528a59397 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "0eb943ae-462c-40dd-a870-7a884ea2db74" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.305s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.933826] env[62965]: DEBUG nova.scheduler.client.report [None req-718b7f9d-fee5-4368-8c0b-8ca81b717274 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 934.987134] env[62965]: DEBUG oslo_concurrency.lockutils [None req-efb9bde1-567b-42dc-9fbf-ca9d2ea5c52e tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "1be808db-d619-4c56-8092-ca52809ec10e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.071s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.987317] env[62965]: DEBUG oslo_vmware.api [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390077, 'name': Destroy_Task, 'duration_secs': 0.36609} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.987736] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Destroyed the VM [ 934.988216] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Deleting Snapshot of the VM instance {{(pid=62965) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 934.988487] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-7aa017c3-06e9-4f7f-8fd5-9036e797e5fd {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.996799] env[62965]: DEBUG oslo_vmware.api [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 934.996799] env[62965]: value = "task-1390078" [ 934.996799] env[62965]: _type = "Task" [ 934.996799] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.006202] env[62965]: DEBUG oslo_vmware.api [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390078, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.249446] env[62965]: DEBUG nova.network.neutron [None req-48f705ca-e256-4c4e-a632-d7db59b218b0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Successfully updated port: 9c46a8bf-6b23-45d3-95a4-f0c751895c71 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 935.438979] env[62965]: DEBUG oslo_concurrency.lockutils [None req-718b7f9d-fee5-4368-8c0b-8ca81b717274 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.827s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.442218] env[62965]: DEBUG oslo_concurrency.lockutils [None req-056ceac3-13ae-4a53-9d5d-071903fb9717 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.407s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.442875] env[62965]: DEBUG nova.objects.instance [None req-056ceac3-13ae-4a53-9d5d-071903fb9717 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lazy-loading 'resources' on Instance uuid d4d9b202-3aa7-4b66-949d-4a2eb34b3adf {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 935.445753] env[62965]: DEBUG oslo_concurrency.lockutils [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquiring lock "a0e9cc1a-bc95-4a44-99c5-aabf85a373f4" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.446066] env[62965]: DEBUG oslo_concurrency.lockutils [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lock "a0e9cc1a-bc95-4a44-99c5-aabf85a373f4" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.446179] env[62965]: INFO nova.compute.manager [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Shelving [ 935.455904] env[62965]: DEBUG nova.compute.manager [req-2172b532-0adc-42ea-9a54-c7dc4836c9d6 req-9297ee58-519e-4df4-b2a1-89a69d7c30fb service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Received event network-vif-plugged-9c46a8bf-6b23-45d3-95a4-f0c751895c71 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 935.456215] env[62965]: DEBUG oslo_concurrency.lockutils [req-2172b532-0adc-42ea-9a54-c7dc4836c9d6 req-9297ee58-519e-4df4-b2a1-89a69d7c30fb service nova] Acquiring lock "d038608c-9264-4fd5-b9ba-0653c98ec0e4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.456371] env[62965]: DEBUG oslo_concurrency.lockutils [req-2172b532-0adc-42ea-9a54-c7dc4836c9d6 req-9297ee58-519e-4df4-b2a1-89a69d7c30fb service nova] Lock "d038608c-9264-4fd5-b9ba-0653c98ec0e4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.456557] env[62965]: DEBUG oslo_concurrency.lockutils [req-2172b532-0adc-42ea-9a54-c7dc4836c9d6 req-9297ee58-519e-4df4-b2a1-89a69d7c30fb service nova] Lock "d038608c-9264-4fd5-b9ba-0653c98ec0e4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.456767] env[62965]: DEBUG nova.compute.manager [req-2172b532-0adc-42ea-9a54-c7dc4836c9d6 req-9297ee58-519e-4df4-b2a1-89a69d7c30fb service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] No waiting events found dispatching network-vif-plugged-9c46a8bf-6b23-45d3-95a4-f0c751895c71 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 935.456916] env[62965]: WARNING nova.compute.manager [req-2172b532-0adc-42ea-9a54-c7dc4836c9d6 req-9297ee58-519e-4df4-b2a1-89a69d7c30fb service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Received unexpected event network-vif-plugged-9c46a8bf-6b23-45d3-95a4-f0c751895c71 for instance with vm_state active and task_state None. [ 935.472754] env[62965]: INFO nova.scheduler.client.report [None req-718b7f9d-fee5-4368-8c0b-8ca81b717274 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Deleted allocations for instance 6def4a38-d3b8-4724-97a1-1859b239d900 [ 935.508780] env[62965]: DEBUG oslo_vmware.api [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390078, 'name': RemoveSnapshot_Task, 'duration_secs': 0.455665} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.509403] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Deleted Snapshot of the VM instance {{(pid=62965) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 935.509861] env[62965]: DEBUG nova.compute.manager [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 935.510869] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afea2083-aa69-4183-b2a1-2c569d90c8da {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.631649] env[62965]: DEBUG nova.compute.manager [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 935.658036] env[62965]: DEBUG nova.virt.hardware [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 935.658036] env[62965]: DEBUG nova.virt.hardware [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 935.658036] env[62965]: DEBUG nova.virt.hardware [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 935.658036] env[62965]: DEBUG nova.virt.hardware [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 935.658328] env[62965]: DEBUG nova.virt.hardware [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 935.659073] env[62965]: DEBUG nova.virt.hardware [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 935.659595] env[62965]: DEBUG nova.virt.hardware [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 935.660648] env[62965]: DEBUG nova.virt.hardware [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 935.660648] env[62965]: DEBUG nova.virt.hardware [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 935.660867] env[62965]: DEBUG nova.virt.hardware [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 935.661051] env[62965]: DEBUG nova.virt.hardware [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 935.662029] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b85ae58e-a1d4-4ec4-94e9-5d601e6b2887 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.672101] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa49500d-4144-4599-b978-13946e4702ca {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.714704] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "fbbcbe60-92f3-497d-b190-520a2547b3a9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.715882] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "fbbcbe60-92f3-497d-b190-520a2547b3a9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.715882] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "fbbcbe60-92f3-497d-b190-520a2547b3a9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.752422] env[62965]: DEBUG oslo_concurrency.lockutils [None req-48f705ca-e256-4c4e-a632-d7db59b218b0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "refresh_cache-d038608c-9264-4fd5-b9ba-0653c98ec0e4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.752562] env[62965]: DEBUG oslo_concurrency.lockutils [None req-48f705ca-e256-4c4e-a632-d7db59b218b0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquired lock "refresh_cache-d038608c-9264-4fd5-b9ba-0653c98ec0e4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.752758] env[62965]: DEBUG nova.network.neutron [None req-48f705ca-e256-4c4e-a632-d7db59b218b0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 935.832077] env[62965]: DEBUG nova.network.neutron [-] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.983030] env[62965]: DEBUG oslo_concurrency.lockutils [None req-718b7f9d-fee5-4368-8c0b-8ca81b717274 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "6def4a38-d3b8-4724-97a1-1859b239d900" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.519s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.025499] env[62965]: INFO nova.compute.manager [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Shelve offloading [ 936.145808] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c83fc68f-a8d3-43cd-909d-2288d5f8b048 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.154879] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56c16375-c9ac-47de-bb29-b743ad639bc9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.190899] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-187cb6eb-5299-4c32-a36b-06852a459b54 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.199831] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73f9b6a3-51a0-4ce6-a24a-88533e0996f9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.214779] env[62965]: DEBUG nova.compute.provider_tree [None req-056ceac3-13ae-4a53-9d5d-071903fb9717 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 936.298309] env[62965]: WARNING nova.network.neutron [None req-48f705ca-e256-4c4e-a632-d7db59b218b0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] a9699d8b-5e76-4666-a236-33ab8f22345e already exists in list: networks containing: ['a9699d8b-5e76-4666-a236-33ab8f22345e']. ignoring it [ 936.334923] env[62965]: INFO nova.compute.manager [-] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Took 1.84 seconds to deallocate network for instance. [ 936.457131] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 936.457131] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-10dccab4-57f5-43fc-9fab-2a6a19406bf1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.463026] env[62965]: DEBUG nova.network.neutron [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Successfully updated port: 212a2219-fc24-44bb-8340-6ad61740edf0 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 936.471487] env[62965]: DEBUG oslo_vmware.api [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 936.471487] env[62965]: value = "task-1390079" [ 936.471487] env[62965]: _type = "Task" [ 936.471487] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.483030] env[62965]: DEBUG oslo_vmware.api [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390079, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.496447] env[62965]: DEBUG oslo_concurrency.lockutils [None req-0ed4cac0-5b86-449d-b75b-6ea9aaa3765c tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "1be808db-d619-4c56-8092-ca52809ec10e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.496805] env[62965]: DEBUG oslo_concurrency.lockutils [None req-0ed4cac0-5b86-449d-b75b-6ea9aaa3765c tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "1be808db-d619-4c56-8092-ca52809ec10e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.497391] env[62965]: DEBUG oslo_concurrency.lockutils [None req-0ed4cac0-5b86-449d-b75b-6ea9aaa3765c tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "1be808db-d619-4c56-8092-ca52809ec10e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.497391] env[62965]: DEBUG oslo_concurrency.lockutils [None req-0ed4cac0-5b86-449d-b75b-6ea9aaa3765c tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "1be808db-d619-4c56-8092-ca52809ec10e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.497391] env[62965]: DEBUG oslo_concurrency.lockutils [None req-0ed4cac0-5b86-449d-b75b-6ea9aaa3765c tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "1be808db-d619-4c56-8092-ca52809ec10e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.499886] env[62965]: INFO nova.compute.manager [None req-0ed4cac0-5b86-449d-b75b-6ea9aaa3765c tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Terminating instance [ 936.533969] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 936.534308] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-26f88f1e-1c48-44ab-8f8c-e6ad828ad544 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.543741] env[62965]: DEBUG oslo_vmware.api [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 936.543741] env[62965]: value = "task-1390080" [ 936.543741] env[62965]: _type = "Task" [ 936.543741] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.555852] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] VM already powered off {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 936.556100] env[62965]: DEBUG nova.compute.manager [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 936.556910] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59a8b43b-0cbc-480c-a702-863ac1aa613a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.564704] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquiring lock "refresh_cache-b59d6023-a047-4b32-bd18-1c7cc05f67d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.564896] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquired lock "refresh_cache-b59d6023-a047-4b32-bd18-1c7cc05f67d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.565084] env[62965]: DEBUG nova.network.neutron [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 936.659756] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20a96db8-5668-4ec2-8934-9f36f0fb13e0 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "0eb943ae-462c-40dd-a870-7a884ea2db74" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.660220] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20a96db8-5668-4ec2-8934-9f36f0fb13e0 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "0eb943ae-462c-40dd-a870-7a884ea2db74" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.718143] env[62965]: DEBUG nova.scheduler.client.report [None req-056ceac3-13ae-4a53-9d5d-071903fb9717 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 936.784023] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "refresh_cache-fbbcbe60-92f3-497d-b190-520a2547b3a9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.784237] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquired lock "refresh_cache-fbbcbe60-92f3-497d-b190-520a2547b3a9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.784424] env[62965]: DEBUG nova.network.neutron [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 936.789145] env[62965]: DEBUG nova.network.neutron [None req-48f705ca-e256-4c4e-a632-d7db59b218b0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Updating instance_info_cache with network_info: [{"id": "b1718061-edf1-4077-b57f-0e54aea365a6", "address": "fa:16:3e:1f:27:16", "network": {"id": "a9699d8b-5e76-4666-a236-33ab8f22345e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1329540850-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1952dedf340f4b7ba0d57af6bba9a749", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a316376e-2ef0-4b1e-b40c-10321ebd7e1a", "external-id": "nsx-vlan-transportzone-942", "segmentation_id": 942, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1718061-ed", "ovs_interfaceid": "b1718061-edf1-4077-b57f-0e54aea365a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "9c46a8bf-6b23-45d3-95a4-f0c751895c71", "address": "fa:16:3e:36:f8:d3", "network": {"id": "a9699d8b-5e76-4666-a236-33ab8f22345e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1329540850-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1952dedf340f4b7ba0d57af6bba9a749", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a316376e-2ef0-4b1e-b40c-10321ebd7e1a", "external-id": "nsx-vlan-transportzone-942", "segmentation_id": 942, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c46a8bf-6b", "ovs_interfaceid": "9c46a8bf-6b23-45d3-95a4-f0c751895c71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.841678] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.965305] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquiring lock "refresh_cache-97954de1-7b4b-4615-8d0f-0e8580a515ed" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.965547] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquired lock "refresh_cache-97954de1-7b4b-4615-8d0f-0e8580a515ed" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.965731] env[62965]: DEBUG nova.network.neutron [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 936.990607] env[62965]: DEBUG oslo_vmware.api [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390079, 'name': PowerOffVM_Task, 'duration_secs': 0.323284} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.991399] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 936.993039] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f912a6b5-0a7e-4051-a6cc-27e9a0dc1e5c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.014123] env[62965]: DEBUG nova.compute.manager [None req-0ed4cac0-5b86-449d-b75b-6ea9aaa3765c tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 937.014325] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-0ed4cac0-5b86-449d-b75b-6ea9aaa3765c tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 937.015664] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17658f55-1c41-40ee-99d4-2f456f4cdf5e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.020831] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0009af73-f3bc-458f-b4fc-6822c8733634 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.034786] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ed4cac0-5b86-449d-b75b-6ea9aaa3765c tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 937.035245] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b6d47877-43b2-4056-ac76-f9e3d2c99465 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.042646] env[62965]: DEBUG oslo_vmware.api [None req-0ed4cac0-5b86-449d-b75b-6ea9aaa3765c tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 937.042646] env[62965]: value = "task-1390081" [ 937.042646] env[62965]: _type = "Task" [ 937.042646] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.055380] env[62965]: DEBUG oslo_vmware.api [None req-0ed4cac0-5b86-449d-b75b-6ea9aaa3765c tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390081, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.167438] env[62965]: DEBUG nova.compute.utils [None req-20a96db8-5668-4ec2-8934-9f36f0fb13e0 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 937.228187] env[62965]: DEBUG oslo_concurrency.lockutils [None req-056ceac3-13ae-4a53-9d5d-071903fb9717 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.786s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.230518] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.979s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.232100] env[62965]: INFO nova.compute.claims [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 937.250014] env[62965]: INFO nova.scheduler.client.report [None req-056ceac3-13ae-4a53-9d5d-071903fb9717 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Deleted allocations for instance d4d9b202-3aa7-4b66-949d-4a2eb34b3adf [ 937.291621] env[62965]: DEBUG oslo_concurrency.lockutils [None req-48f705ca-e256-4c4e-a632-d7db59b218b0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Releasing lock "refresh_cache-d038608c-9264-4fd5-b9ba-0653c98ec0e4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.292325] env[62965]: DEBUG oslo_concurrency.lockutils [None req-48f705ca-e256-4c4e-a632-d7db59b218b0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "d038608c-9264-4fd5-b9ba-0653c98ec0e4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.292498] env[62965]: DEBUG oslo_concurrency.lockutils [None req-48f705ca-e256-4c4e-a632-d7db59b218b0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquired lock "d038608c-9264-4fd5-b9ba-0653c98ec0e4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.293584] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80a2184e-ea22-4f42-8a24-8b12845473b9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.312530] env[62965]: DEBUG nova.virt.hardware [None req-48f705ca-e256-4c4e-a632-d7db59b218b0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 937.312697] env[62965]: DEBUG nova.virt.hardware [None req-48f705ca-e256-4c4e-a632-d7db59b218b0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 937.312806] env[62965]: DEBUG nova.virt.hardware [None req-48f705ca-e256-4c4e-a632-d7db59b218b0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 937.312931] env[62965]: DEBUG nova.virt.hardware [None req-48f705ca-e256-4c4e-a632-d7db59b218b0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 937.313108] env[62965]: DEBUG nova.virt.hardware [None req-48f705ca-e256-4c4e-a632-d7db59b218b0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 937.313265] env[62965]: DEBUG nova.virt.hardware [None req-48f705ca-e256-4c4e-a632-d7db59b218b0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 937.313475] env[62965]: DEBUG nova.virt.hardware [None req-48f705ca-e256-4c4e-a632-d7db59b218b0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 937.313635] env[62965]: DEBUG nova.virt.hardware [None req-48f705ca-e256-4c4e-a632-d7db59b218b0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 937.313807] env[62965]: DEBUG nova.virt.hardware [None req-48f705ca-e256-4c4e-a632-d7db59b218b0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 937.314117] env[62965]: DEBUG nova.virt.hardware [None req-48f705ca-e256-4c4e-a632-d7db59b218b0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 937.314194] env[62965]: DEBUG nova.virt.hardware [None req-48f705ca-e256-4c4e-a632-d7db59b218b0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 937.320419] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-48f705ca-e256-4c4e-a632-d7db59b218b0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Reconfiguring VM to attach interface {{(pid=62965) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 937.321473] env[62965]: DEBUG nova.network.neutron [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Updating instance_info_cache with network_info: [{"id": "d2d44211-6856-42ed-8f9f-3b1a58b31349", "address": "fa:16:3e:3b:f0:e9", "network": {"id": "49dd42b0-6397-4fdd-a319-ed920c72a5af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-373172566-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.230", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d91c99b14018469e8998fd1ccc0fab4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2d44211-68", "ovs_interfaceid": "d2d44211-6856-42ed-8f9f-3b1a58b31349", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.322775] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd351e21-5512-40c0-91de-8611787c4b6b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.337692] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Releasing lock "refresh_cache-b59d6023-a047-4b32-bd18-1c7cc05f67d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.345775] env[62965]: DEBUG oslo_vmware.api [None req-48f705ca-e256-4c4e-a632-d7db59b218b0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Waiting for the task: (returnval){ [ 937.345775] env[62965]: value = "task-1390082" [ 937.345775] env[62965]: _type = "Task" [ 937.345775] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.354792] env[62965]: DEBUG oslo_vmware.api [None req-48f705ca-e256-4c4e-a632-d7db59b218b0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390082, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.535260] env[62965]: DEBUG nova.network.neutron [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 937.537897] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Creating Snapshot of the VM instance {{(pid=62965) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 937.538245] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-417adc9b-b61e-42d9-9e1e-de81dc5a6ba0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.550690] env[62965]: DEBUG oslo_vmware.api [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 937.550690] env[62965]: value = "task-1390083" [ 937.550690] env[62965]: _type = "Task" [ 937.550690] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.557926] env[62965]: DEBUG oslo_vmware.api [None req-0ed4cac0-5b86-449d-b75b-6ea9aaa3765c tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390081, 'name': PowerOffVM_Task, 'duration_secs': 0.178306} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.558671] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ed4cac0-5b86-449d-b75b-6ea9aaa3765c tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 937.558923] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-0ed4cac0-5b86-449d-b75b-6ea9aaa3765c tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 937.559207] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5cd8efdc-6ff7-45ed-b9fe-43d9af41312d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.563974] env[62965]: DEBUG oslo_vmware.api [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390083, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.646661] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-0ed4cac0-5b86-449d-b75b-6ea9aaa3765c tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 937.646895] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-0ed4cac0-5b86-449d-b75b-6ea9aaa3765c tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Deleting contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 937.647102] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ed4cac0-5b86-449d-b75b-6ea9aaa3765c tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Deleting the datastore file [datastore2] 1be808db-d619-4c56-8092-ca52809ec10e {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 937.647409] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2718257b-3965-4d54-b2c2-992681249213 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.654531] env[62965]: DEBUG oslo_vmware.api [None req-0ed4cac0-5b86-449d-b75b-6ea9aaa3765c tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 937.654531] env[62965]: value = "task-1390085" [ 937.654531] env[62965]: _type = "Task" [ 937.654531] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.663314] env[62965]: DEBUG oslo_vmware.api [None req-0ed4cac0-5b86-449d-b75b-6ea9aaa3765c tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390085, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.672101] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20a96db8-5668-4ec2-8934-9f36f0fb13e0 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "0eb943ae-462c-40dd-a870-7a884ea2db74" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.012s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.723797] env[62965]: DEBUG nova.network.neutron [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Updating instance_info_cache with network_info: [{"id": "3edb59ed-a2b8-4d31-b975-9c9d64c3a22b", "address": "fa:16:3e:92:26:4b", "network": {"id": "a5246cc1-6001-4c20-9724-fd36a0437582", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-644702631-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bfa5cfdc7d5496482d87675f2a122dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3edb59ed-a2", "ovs_interfaceid": "3edb59ed-a2b8-4d31-b975-9c9d64c3a22b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.758238] env[62965]: DEBUG oslo_concurrency.lockutils [None req-056ceac3-13ae-4a53-9d5d-071903fb9717 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "d4d9b202-3aa7-4b66-949d-4a2eb34b3adf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.193s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.807937] env[62965]: DEBUG nova.network.neutron [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Updating instance_info_cache with network_info: [{"id": "212a2219-fc24-44bb-8340-6ad61740edf0", "address": "fa:16:3e:ba:ed:e3", "network": {"id": "51f79fa0-8980-40d2-9255-a8da5b8fc5d7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-909239894-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1c6e8403409438f9ff10235cd2533bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap212a2219-fc", "ovs_interfaceid": "212a2219-fc24-44bb-8340-6ad61740edf0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.857942] env[62965]: DEBUG oslo_vmware.api [None req-48f705ca-e256-4c4e-a632-d7db59b218b0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390082, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.953461] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 937.954375] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c28fa55-b2fa-42f3-a34e-1a145781afba {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.963398] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 937.963669] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f9e33737-4110-4ff2-90eb-106a18d3630a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.062549] env[62965]: DEBUG oslo_vmware.api [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390083, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.064152] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 938.064387] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Deleting contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 938.064586] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Deleting the datastore file [datastore1] b59d6023-a047-4b32-bd18-1c7cc05f67d7 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 938.064851] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4ba820a8-73d6-4f80-bd6d-d6bcb1f87c29 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.072071] env[62965]: DEBUG oslo_vmware.api [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 938.072071] env[62965]: value = "task-1390087" [ 938.072071] env[62965]: _type = "Task" [ 938.072071] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.080787] env[62965]: DEBUG oslo_vmware.api [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390087, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.166469] env[62965]: DEBUG oslo_vmware.api [None req-0ed4cac0-5b86-449d-b75b-6ea9aaa3765c tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390085, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.231603} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.166818] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ed4cac0-5b86-449d-b75b-6ea9aaa3765c tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 938.167094] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-0ed4cac0-5b86-449d-b75b-6ea9aaa3765c tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Deleted contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 938.167332] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-0ed4cac0-5b86-449d-b75b-6ea9aaa3765c tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 938.168463] env[62965]: INFO nova.compute.manager [None req-0ed4cac0-5b86-449d-b75b-6ea9aaa3765c tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Took 1.15 seconds to destroy the instance on the hypervisor. [ 938.168463] env[62965]: DEBUG oslo.service.loopingcall [None req-0ed4cac0-5b86-449d-b75b-6ea9aaa3765c tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 938.168463] env[62965]: DEBUG nova.compute.manager [-] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 938.168463] env[62965]: DEBUG nova.network.neutron [-] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 938.226426] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Releasing lock "refresh_cache-fbbcbe60-92f3-497d-b190-520a2547b3a9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.311257] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Releasing lock "refresh_cache-97954de1-7b4b-4615-8d0f-0e8580a515ed" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.311568] env[62965]: DEBUG nova.compute.manager [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Instance network_info: |[{"id": "212a2219-fc24-44bb-8340-6ad61740edf0", "address": "fa:16:3e:ba:ed:e3", "network": {"id": "51f79fa0-8980-40d2-9255-a8da5b8fc5d7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-909239894-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1c6e8403409438f9ff10235cd2533bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap212a2219-fc", "ovs_interfaceid": "212a2219-fc24-44bb-8340-6ad61740edf0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 938.312246] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ba:ed:e3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c6934071-bf85-4591-9c7d-55c7ea131262', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '212a2219-fc24-44bb-8340-6ad61740edf0', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 938.320950] env[62965]: DEBUG oslo.service.loopingcall [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 938.321127] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 938.322051] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-480799fc-844a-44c3-8605-2a6fa2fbe13d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.348025] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 938.348025] env[62965]: value = "task-1390088" [ 938.348025] env[62965]: _type = "Task" [ 938.348025] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.361511] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390088, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.366905] env[62965]: DEBUG oslo_vmware.api [None req-48f705ca-e256-4c4e-a632-d7db59b218b0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390082, 'name': ReconfigVM_Task, 'duration_secs': 0.673772} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.367883] env[62965]: DEBUG oslo_concurrency.lockutils [None req-48f705ca-e256-4c4e-a632-d7db59b218b0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Releasing lock "d038608c-9264-4fd5-b9ba-0653c98ec0e4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.368367] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-48f705ca-e256-4c4e-a632-d7db59b218b0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Reconfigured VM to attach interface {{(pid=62965) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 938.487075] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3aa0d4d-9874-41fe-b0cc-ad805e384111 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.495578] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11d77eaa-aa66-4320-9f06-51a6572501c0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.529099] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fa47af5-1d38-47e6-b86c-c9064d55610e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.537818] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad2a3d77-4d20-406e-bca9-eb9c40b08589 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.552189] env[62965]: DEBUG nova.compute.provider_tree [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 938.565096] env[62965]: DEBUG oslo_vmware.api [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390083, 'name': CreateSnapshot_Task, 'duration_secs': 0.869886} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.565096] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Created Snapshot of the VM instance {{(pid=62965) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 938.565096] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9740fd4-8ffd-4144-9a80-597ef2d1a391 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.586545] env[62965]: DEBUG oslo_vmware.api [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390087, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.192068} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.586804] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 938.587029] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Deleted contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 938.587183] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 938.606820] env[62965]: INFO nova.scheduler.client.report [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Deleted allocations for instance b59d6023-a047-4b32-bd18-1c7cc05f67d7 [ 938.742961] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20a96db8-5668-4ec2-8934-9f36f0fb13e0 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "0eb943ae-462c-40dd-a870-7a884ea2db74" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.743312] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20a96db8-5668-4ec2-8934-9f36f0fb13e0 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "0eb943ae-462c-40dd-a870-7a884ea2db74" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.743552] env[62965]: INFO nova.compute.manager [None req-20a96db8-5668-4ec2-8934-9f36f0fb13e0 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Attaching volume faaca442-1091-4470-8417-2b2608e503c8 to /dev/sdb [ 938.752078] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bb52cf1-27be-4562-9a97-d33b4c92ae8a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.773674] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-473016f9-8f4b-4545-84c6-c8b7218e02eb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.781894] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Updating instance 'fbbcbe60-92f3-497d-b190-520a2547b3a9' progress to 83 {{(pid=62965) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 938.787601] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b40adac-01fa-4aba-a907-980528f6cca0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.795344] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42dba350-8b31-4ec0-9103-ed188b7fcbc3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.810628] env[62965]: DEBUG nova.virt.block_device [None req-20a96db8-5668-4ec2-8934-9f36f0fb13e0 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Updating existing volume attachment record: ec49ec54-a5fe-46c3-b2a3-e4e4199b0439 {{(pid=62965) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 938.855662] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390088, 'name': CreateVM_Task, 'duration_secs': 0.485425} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.855834] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 938.856568] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.856823] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.857131] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 938.857404] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d930839-ad9d-413b-908d-d7f997ff43bd {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.863202] env[62965]: DEBUG oslo_vmware.api [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 938.863202] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52a3a7b6-f4f3-03a1-5fb5-2eb542ac84b3" [ 938.863202] env[62965]: _type = "Task" [ 938.863202] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.871449] env[62965]: DEBUG oslo_vmware.api [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52a3a7b6-f4f3-03a1-5fb5-2eb542ac84b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.873643] env[62965]: DEBUG oslo_concurrency.lockutils [None req-48f705ca-e256-4c4e-a632-d7db59b218b0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "interface-d038608c-9264-4fd5-b9ba-0653c98ec0e4-9c46a8bf-6b23-45d3-95a4-f0c751895c71" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.691s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.940409] env[62965]: DEBUG nova.network.neutron [-] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.055704] env[62965]: DEBUG nova.scheduler.client.report [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 939.085144] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Creating linked-clone VM from snapshot {{(pid=62965) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 939.085144] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-bcfd48c2-4928-4576-a94c-3be54b554688 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.094596] env[62965]: DEBUG oslo_vmware.api [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 939.094596] env[62965]: value = "task-1390091" [ 939.094596] env[62965]: _type = "Task" [ 939.094596] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.109571] env[62965]: DEBUG oslo_vmware.api [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390091, 'name': CloneVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.111825] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.122164] env[62965]: DEBUG nova.compute.manager [req-4ded6c47-6c91-4cc1-a02c-bd51803361b0 req-00b1e0d7-f043-445b-9b41-d08806ffe78a service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Received event network-changed-9c46a8bf-6b23-45d3-95a4-f0c751895c71 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 939.122385] env[62965]: DEBUG nova.compute.manager [req-4ded6c47-6c91-4cc1-a02c-bd51803361b0 req-00b1e0d7-f043-445b-9b41-d08806ffe78a service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Refreshing instance network info cache due to event network-changed-9c46a8bf-6b23-45d3-95a4-f0c751895c71. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 939.122936] env[62965]: DEBUG oslo_concurrency.lockutils [req-4ded6c47-6c91-4cc1-a02c-bd51803361b0 req-00b1e0d7-f043-445b-9b41-d08806ffe78a service nova] Acquiring lock "refresh_cache-d038608c-9264-4fd5-b9ba-0653c98ec0e4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.123119] env[62965]: DEBUG oslo_concurrency.lockutils [req-4ded6c47-6c91-4cc1-a02c-bd51803361b0 req-00b1e0d7-f043-445b-9b41-d08806ffe78a service nova] Acquired lock "refresh_cache-d038608c-9264-4fd5-b9ba-0653c98ec0e4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.123298] env[62965]: DEBUG nova.network.neutron [req-4ded6c47-6c91-4cc1-a02c-bd51803361b0 req-00b1e0d7-f043-445b-9b41-d08806ffe78a service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Refreshing network info cache for port 9c46a8bf-6b23-45d3-95a4-f0c751895c71 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 939.293172] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 939.293511] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-edff367b-fbd1-4aff-8477-5b6ded454bfa {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.303813] env[62965]: DEBUG oslo_vmware.api [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 939.303813] env[62965]: value = "task-1390093" [ 939.303813] env[62965]: _type = "Task" [ 939.303813] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.312920] env[62965]: DEBUG oslo_vmware.api [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390093, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.378972] env[62965]: DEBUG oslo_vmware.api [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52a3a7b6-f4f3-03a1-5fb5-2eb542ac84b3, 'name': SearchDatastore_Task, 'duration_secs': 0.0199} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.379683] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.379930] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 939.380364] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.380720] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.380917] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 939.381470] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0b32df77-f35a-44a9-a4d2-08eac3003179 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.391409] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 939.391540] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 939.392359] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3511a992-1dd4-4313-ae50-9780db517904 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.399761] env[62965]: DEBUG oslo_vmware.api [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 939.399761] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5267db79-ee8f-d3da-7b21-d6767753fff0" [ 939.399761] env[62965]: _type = "Task" [ 939.399761] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.409565] env[62965]: DEBUG oslo_vmware.api [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5267db79-ee8f-d3da-7b21-d6767753fff0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.443366] env[62965]: INFO nova.compute.manager [-] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Took 1.27 seconds to deallocate network for instance. [ 939.562624] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.331s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.562624] env[62965]: DEBUG nova.compute.manager [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 939.564667] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.723s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.564915] env[62965]: DEBUG nova.objects.instance [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Lazy-loading 'resources' on Instance uuid 64cf898c-2c97-4daa-80c8-0a1c252cbcf2 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 939.606187] env[62965]: DEBUG oslo_vmware.api [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390091, 'name': CloneVM_Task} progress is 94%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.684328] env[62965]: DEBUG oslo_concurrency.lockutils [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquiring lock "a63b301d-8d52-4fec-951f-a15cda9449ac" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.684532] env[62965]: DEBUG oslo_concurrency.lockutils [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "a63b301d-8d52-4fec-951f-a15cda9449ac" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.815384] env[62965]: DEBUG oslo_vmware.api [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390093, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.864735] env[62965]: DEBUG nova.network.neutron [req-4ded6c47-6c91-4cc1-a02c-bd51803361b0 req-00b1e0d7-f043-445b-9b41-d08806ffe78a service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Updated VIF entry in instance network info cache for port 9c46a8bf-6b23-45d3-95a4-f0c751895c71. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 939.865429] env[62965]: DEBUG nova.network.neutron [req-4ded6c47-6c91-4cc1-a02c-bd51803361b0 req-00b1e0d7-f043-445b-9b41-d08806ffe78a service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Updating instance_info_cache with network_info: [{"id": "b1718061-edf1-4077-b57f-0e54aea365a6", "address": "fa:16:3e:1f:27:16", "network": {"id": "a9699d8b-5e76-4666-a236-33ab8f22345e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1329540850-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1952dedf340f4b7ba0d57af6bba9a749", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a316376e-2ef0-4b1e-b40c-10321ebd7e1a", "external-id": "nsx-vlan-transportzone-942", "segmentation_id": 942, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1718061-ed", "ovs_interfaceid": "b1718061-edf1-4077-b57f-0e54aea365a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "9c46a8bf-6b23-45d3-95a4-f0c751895c71", "address": "fa:16:3e:36:f8:d3", "network": {"id": "a9699d8b-5e76-4666-a236-33ab8f22345e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1329540850-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1952dedf340f4b7ba0d57af6bba9a749", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a316376e-2ef0-4b1e-b40c-10321ebd7e1a", "external-id": "nsx-vlan-transportzone-942", "segmentation_id": 942, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c46a8bf-6b", "ovs_interfaceid": "9c46a8bf-6b23-45d3-95a4-f0c751895c71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.912359] env[62965]: DEBUG oslo_vmware.api [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5267db79-ee8f-d3da-7b21-d6767753fff0, 'name': SearchDatastore_Task, 'duration_secs': 0.011845} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.913197] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b39603d3-08af-419e-ae2f-cf73f694ef79 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.920223] env[62965]: DEBUG oslo_vmware.api [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 939.920223] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52f28726-0b0c-e39a-2de6-3470b8821b42" [ 939.920223] env[62965]: _type = "Task" [ 939.920223] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.928861] env[62965]: DEBUG oslo_vmware.api [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52f28726-0b0c-e39a-2de6-3470b8821b42, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.951159] env[62965]: DEBUG oslo_concurrency.lockutils [None req-0ed4cac0-5b86-449d-b75b-6ea9aaa3765c tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.068198] env[62965]: DEBUG nova.compute.utils [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 940.072758] env[62965]: DEBUG nova.compute.manager [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 940.072926] env[62965]: DEBUG nova.network.neutron [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 940.107869] env[62965]: DEBUG oslo_vmware.api [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390091, 'name': CloneVM_Task} progress is 94%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.120623] env[62965]: DEBUG nova.policy [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '736eaca6d3584f74856c61ff0e6ce03e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd3edf0643c9d40e9bb568742b7a7a508', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 940.188278] env[62965]: DEBUG nova.compute.manager [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 940.269254] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efcf50d7-be73-47ab-8b9e-98914882b0f9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.277723] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-993eb4ce-911a-4bc5-987e-ce9b6556ea1c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.312782] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7992afb-bc43-4843-b9a2-c6f286d8eff7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.321109] env[62965]: DEBUG oslo_vmware.api [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390093, 'name': PowerOnVM_Task, 'duration_secs': 0.63031} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.323276] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 940.323482] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7c832d2b-9e1f-4e1d-bfd9-422c09f18cd1 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Updating instance 'fbbcbe60-92f3-497d-b190-520a2547b3a9' progress to 100 {{(pid=62965) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 940.328099] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22bc2f05-df85-4ed5-bf3d-6425faca02e4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.342871] env[62965]: DEBUG nova.compute.provider_tree [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 940.371785] env[62965]: DEBUG oslo_concurrency.lockutils [req-4ded6c47-6c91-4cc1-a02c-bd51803361b0 req-00b1e0d7-f043-445b-9b41-d08806ffe78a service nova] Releasing lock "refresh_cache-d038608c-9264-4fd5-b9ba-0653c98ec0e4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.371785] env[62965]: DEBUG nova.compute.manager [req-4ded6c47-6c91-4cc1-a02c-bd51803361b0 req-00b1e0d7-f043-445b-9b41-d08806ffe78a service nova] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Received event network-vif-deleted-97930362-17f2-40a5-bc09-a9df5f010935 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 940.371785] env[62965]: DEBUG nova.compute.manager [req-4ded6c47-6c91-4cc1-a02c-bd51803361b0 req-00b1e0d7-f043-445b-9b41-d08806ffe78a service nova] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Received event network-vif-plugged-212a2219-fc24-44bb-8340-6ad61740edf0 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 940.371785] env[62965]: DEBUG oslo_concurrency.lockutils [req-4ded6c47-6c91-4cc1-a02c-bd51803361b0 req-00b1e0d7-f043-445b-9b41-d08806ffe78a service nova] Acquiring lock "97954de1-7b4b-4615-8d0f-0e8580a515ed-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.371785] env[62965]: DEBUG oslo_concurrency.lockutils [req-4ded6c47-6c91-4cc1-a02c-bd51803361b0 req-00b1e0d7-f043-445b-9b41-d08806ffe78a service nova] Lock "97954de1-7b4b-4615-8d0f-0e8580a515ed-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.371785] env[62965]: DEBUG oslo_concurrency.lockutils [req-4ded6c47-6c91-4cc1-a02c-bd51803361b0 req-00b1e0d7-f043-445b-9b41-d08806ffe78a service nova] Lock "97954de1-7b4b-4615-8d0f-0e8580a515ed-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.371785] env[62965]: DEBUG nova.compute.manager [req-4ded6c47-6c91-4cc1-a02c-bd51803361b0 req-00b1e0d7-f043-445b-9b41-d08806ffe78a service nova] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] No waiting events found dispatching network-vif-plugged-212a2219-fc24-44bb-8340-6ad61740edf0 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 940.371785] env[62965]: WARNING nova.compute.manager [req-4ded6c47-6c91-4cc1-a02c-bd51803361b0 req-00b1e0d7-f043-445b-9b41-d08806ffe78a service nova] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Received unexpected event network-vif-plugged-212a2219-fc24-44bb-8340-6ad61740edf0 for instance with vm_state building and task_state spawning. [ 940.371785] env[62965]: DEBUG nova.compute.manager [req-4ded6c47-6c91-4cc1-a02c-bd51803361b0 req-00b1e0d7-f043-445b-9b41-d08806ffe78a service nova] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Received event network-changed-212a2219-fc24-44bb-8340-6ad61740edf0 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 940.372134] env[62965]: DEBUG nova.compute.manager [req-4ded6c47-6c91-4cc1-a02c-bd51803361b0 req-00b1e0d7-f043-445b-9b41-d08806ffe78a service nova] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Refreshing instance network info cache due to event network-changed-212a2219-fc24-44bb-8340-6ad61740edf0. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 940.372134] env[62965]: DEBUG oslo_concurrency.lockutils [req-4ded6c47-6c91-4cc1-a02c-bd51803361b0 req-00b1e0d7-f043-445b-9b41-d08806ffe78a service nova] Acquiring lock "refresh_cache-97954de1-7b4b-4615-8d0f-0e8580a515ed" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.372759] env[62965]: DEBUG oslo_concurrency.lockutils [req-4ded6c47-6c91-4cc1-a02c-bd51803361b0 req-00b1e0d7-f043-445b-9b41-d08806ffe78a service nova] Acquired lock "refresh_cache-97954de1-7b4b-4615-8d0f-0e8580a515ed" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.372759] env[62965]: DEBUG nova.network.neutron [req-4ded6c47-6c91-4cc1-a02c-bd51803361b0 req-00b1e0d7-f043-445b-9b41-d08806ffe78a service nova] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Refreshing network info cache for port 212a2219-fc24-44bb-8340-6ad61740edf0 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 940.385919] env[62965]: DEBUG nova.network.neutron [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Successfully created port: 3eafc53f-b78d-43b3-9341-89b3a9cd0ecb {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 940.435399] env[62965]: DEBUG oslo_vmware.api [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52f28726-0b0c-e39a-2de6-3470b8821b42, 'name': SearchDatastore_Task, 'duration_secs': 0.011219} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.435712] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.435995] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] 97954de1-7b4b-4615-8d0f-0e8580a515ed/97954de1-7b4b-4615-8d0f-0e8580a515ed.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 940.436301] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f2fb9f09-8828-4b2b-89be-1f00f42a5654 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.444100] env[62965]: DEBUG oslo_vmware.api [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 940.444100] env[62965]: value = "task-1390094" [ 940.444100] env[62965]: _type = "Task" [ 940.444100] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.452914] env[62965]: DEBUG oslo_vmware.api [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390094, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.504029] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e73a534f-af98-4393-85f0-5e6378f4083e tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "interface-d038608c-9264-4fd5-b9ba-0653c98ec0e4-9c46a8bf-6b23-45d3-95a4-f0c751895c71" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.504330] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e73a534f-af98-4393-85f0-5e6378f4083e tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "interface-d038608c-9264-4fd5-b9ba-0653c98ec0e4-9c46a8bf-6b23-45d3-95a4-f0c751895c71" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.573304] env[62965]: DEBUG nova.compute.manager [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 940.606949] env[62965]: DEBUG oslo_vmware.api [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390091, 'name': CloneVM_Task, 'duration_secs': 1.330515} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.607285] env[62965]: INFO nova.virt.vmwareapi.vmops [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Created linked-clone VM from snapshot [ 940.608134] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91e1d8e9-a6a5-4e9a-ae34-d6d78fd7aed7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.617181] env[62965]: DEBUG nova.virt.vmwareapi.images [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Uploading image c1230705-5bdc-4c15-a740-13db3313def2 {{(pid=62965) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 940.644468] env[62965]: DEBUG oslo_vmware.rw_handles [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 940.644468] env[62965]: value = "vm-295023" [ 940.644468] env[62965]: _type = "VirtualMachine" [ 940.644468] env[62965]: }. {{(pid=62965) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 940.651442] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-336a4f32-4a60-4e02-8e32-0bef6ec4cd43 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.663655] env[62965]: DEBUG oslo_vmware.rw_handles [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lease: (returnval){ [ 940.663655] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52ee4815-88c1-b492-0f5c-d8e927282183" [ 940.663655] env[62965]: _type = "HttpNfcLease" [ 940.663655] env[62965]: } obtained for exporting VM: (result){ [ 940.663655] env[62965]: value = "vm-295023" [ 940.663655] env[62965]: _type = "VirtualMachine" [ 940.663655] env[62965]: }. {{(pid=62965) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 940.664221] env[62965]: DEBUG oslo_vmware.api [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the lease: (returnval){ [ 940.664221] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52ee4815-88c1-b492-0f5c-d8e927282183" [ 940.664221] env[62965]: _type = "HttpNfcLease" [ 940.664221] env[62965]: } to be ready. {{(pid=62965) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 940.675673] env[62965]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 940.675673] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52ee4815-88c1-b492-0f5c-d8e927282183" [ 940.675673] env[62965]: _type = "HttpNfcLease" [ 940.675673] env[62965]: } is initializing. {{(pid=62965) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 940.713790] env[62965]: DEBUG oslo_concurrency.lockutils [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.863444] env[62965]: ERROR nova.scheduler.client.report [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] [req-a1c8c218-36c6-4e81-8a09-8cb1c1b24e30] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-a1c8c218-36c6-4e81-8a09-8cb1c1b24e30"}]} [ 940.880846] env[62965]: DEBUG nova.scheduler.client.report [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Refreshing inventories for resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 940.897573] env[62965]: DEBUG nova.scheduler.client.report [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Updating ProviderTree inventory for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 940.897920] env[62965]: DEBUG nova.compute.provider_tree [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 940.913650] env[62965]: DEBUG nova.scheduler.client.report [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Refreshing aggregate associations for resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8, aggregates: None {{(pid=62965) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 940.937439] env[62965]: DEBUG nova.scheduler.client.report [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Refreshing trait associations for resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64 {{(pid=62965) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 940.956708] env[62965]: DEBUG oslo_vmware.api [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390094, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.009482] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e73a534f-af98-4393-85f0-5e6378f4083e tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "d038608c-9264-4fd5-b9ba-0653c98ec0e4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.009666] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e73a534f-af98-4393-85f0-5e6378f4083e tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquired lock "d038608c-9264-4fd5-b9ba-0653c98ec0e4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.011239] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e666bc8-5256-4ac1-a2c7-3cac3a07f150 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.037849] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea64fa3a-40cb-41f7-8534-fd64eaf6a2c4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.066105] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e73a534f-af98-4393-85f0-5e6378f4083e tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Reconfiguring VM to detach interface {{(pid=62965) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 941.066809] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-56bc51ee-c7c1-4778-8da4-a6aaa6c1780c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.092756] env[62965]: DEBUG oslo_vmware.api [None req-e73a534f-af98-4393-85f0-5e6378f4083e tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Waiting for the task: (returnval){ [ 941.092756] env[62965]: value = "task-1390096" [ 941.092756] env[62965]: _type = "Task" [ 941.092756] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.101041] env[62965]: DEBUG oslo_vmware.api [None req-e73a534f-af98-4393-85f0-5e6378f4083e tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390096, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.150286] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquiring lock "b59d6023-a047-4b32-bd18-1c7cc05f67d7" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.176153] env[62965]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 941.176153] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52ee4815-88c1-b492-0f5c-d8e927282183" [ 941.176153] env[62965]: _type = "HttpNfcLease" [ 941.176153] env[62965]: } is ready. {{(pid=62965) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 941.176432] env[62965]: DEBUG oslo_vmware.rw_handles [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 941.176432] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52ee4815-88c1-b492-0f5c-d8e927282183" [ 941.176432] env[62965]: _type = "HttpNfcLease" [ 941.176432] env[62965]: }. {{(pid=62965) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 941.176933] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad960973-2a65-4bf6-8840-071fd72160b3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.184491] env[62965]: DEBUG oslo_vmware.rw_handles [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5205ac28-03ec-e1dd-1f2d-02c5b4ff905d/disk-0.vmdk from lease info. {{(pid=62965) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 941.184670] env[62965]: DEBUG oslo_vmware.rw_handles [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5205ac28-03ec-e1dd-1f2d-02c5b4ff905d/disk-0.vmdk for reading. {{(pid=62965) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 941.287539] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94e2780d-2478-4aff-8f7a-75c38622cccf {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.296053] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bab15cd0-d63a-4c95-96b3-449faeea881b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.301704] env[62965]: DEBUG nova.compute.manager [req-b98ff61b-7973-407b-ad0d-7874d8f8f507 req-88d95682-ca7d-4567-858f-bfc35a11cb09 service nova] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Received event network-vif-unplugged-d2d44211-6856-42ed-8f9f-3b1a58b31349 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 941.301928] env[62965]: DEBUG oslo_concurrency.lockutils [req-b98ff61b-7973-407b-ad0d-7874d8f8f507 req-88d95682-ca7d-4567-858f-bfc35a11cb09 service nova] Acquiring lock "b59d6023-a047-4b32-bd18-1c7cc05f67d7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.302155] env[62965]: DEBUG oslo_concurrency.lockutils [req-b98ff61b-7973-407b-ad0d-7874d8f8f507 req-88d95682-ca7d-4567-858f-bfc35a11cb09 service nova] Lock "b59d6023-a047-4b32-bd18-1c7cc05f67d7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.302325] env[62965]: DEBUG oslo_concurrency.lockutils [req-b98ff61b-7973-407b-ad0d-7874d8f8f507 req-88d95682-ca7d-4567-858f-bfc35a11cb09 service nova] Lock "b59d6023-a047-4b32-bd18-1c7cc05f67d7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.302505] env[62965]: DEBUG nova.compute.manager [req-b98ff61b-7973-407b-ad0d-7874d8f8f507 req-88d95682-ca7d-4567-858f-bfc35a11cb09 service nova] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] No waiting events found dispatching network-vif-unplugged-d2d44211-6856-42ed-8f9f-3b1a58b31349 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 941.302674] env[62965]: WARNING nova.compute.manager [req-b98ff61b-7973-407b-ad0d-7874d8f8f507 req-88d95682-ca7d-4567-858f-bfc35a11cb09 service nova] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Received unexpected event network-vif-unplugged-d2d44211-6856-42ed-8f9f-3b1a58b31349 for instance with vm_state shelved_offloaded and task_state unshelving. [ 941.302837] env[62965]: DEBUG nova.compute.manager [req-b98ff61b-7973-407b-ad0d-7874d8f8f507 req-88d95682-ca7d-4567-858f-bfc35a11cb09 service nova] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Received event network-changed-d2d44211-6856-42ed-8f9f-3b1a58b31349 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 941.302994] env[62965]: DEBUG nova.compute.manager [req-b98ff61b-7973-407b-ad0d-7874d8f8f507 req-88d95682-ca7d-4567-858f-bfc35a11cb09 service nova] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Refreshing instance network info cache due to event network-changed-d2d44211-6856-42ed-8f9f-3b1a58b31349. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 941.303197] env[62965]: DEBUG oslo_concurrency.lockutils [req-b98ff61b-7973-407b-ad0d-7874d8f8f507 req-88d95682-ca7d-4567-858f-bfc35a11cb09 service nova] Acquiring lock "refresh_cache-b59d6023-a047-4b32-bd18-1c7cc05f67d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.303335] env[62965]: DEBUG oslo_concurrency.lockutils [req-b98ff61b-7973-407b-ad0d-7874d8f8f507 req-88d95682-ca7d-4567-858f-bfc35a11cb09 service nova] Acquired lock "refresh_cache-b59d6023-a047-4b32-bd18-1c7cc05f67d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.303491] env[62965]: DEBUG nova.network.neutron [req-b98ff61b-7973-407b-ad0d-7874d8f8f507 req-88d95682-ca7d-4567-858f-bfc35a11cb09 service nova] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Refreshing network info cache for port d2d44211-6856-42ed-8f9f-3b1a58b31349 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 941.335064] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f66d110-5ce2-4d3f-b55c-45033af88773 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.349375] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38ec8fef-73fb-4e3b-80da-2dbc05f24c1b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.368452] env[62965]: DEBUG nova.compute.provider_tree [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 941.384389] env[62965]: DEBUG nova.network.neutron [req-4ded6c47-6c91-4cc1-a02c-bd51803361b0 req-00b1e0d7-f043-445b-9b41-d08806ffe78a service nova] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Updated VIF entry in instance network info cache for port 212a2219-fc24-44bb-8340-6ad61740edf0. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 941.384389] env[62965]: DEBUG nova.network.neutron [req-4ded6c47-6c91-4cc1-a02c-bd51803361b0 req-00b1e0d7-f043-445b-9b41-d08806ffe78a service nova] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Updating instance_info_cache with network_info: [{"id": "212a2219-fc24-44bb-8340-6ad61740edf0", "address": "fa:16:3e:ba:ed:e3", "network": {"id": "51f79fa0-8980-40d2-9255-a8da5b8fc5d7", "bridge": "br-int", "label": "tempest-ImagesTestJSON-909239894-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e1c6e8403409438f9ff10235cd2533bc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c6934071-bf85-4591-9c7d-55c7ea131262", "external-id": "nsx-vlan-transportzone-452", "segmentation_id": 452, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap212a2219-fc", "ovs_interfaceid": "212a2219-fc24-44bb-8340-6ad61740edf0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.385288] env[62965]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-fb6521a2-f9a5-4b1c-84af-9aff34c32a1b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.457361] env[62965]: DEBUG oslo_vmware.api [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390094, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.550957} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.457361] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] 97954de1-7b4b-4615-8d0f-0e8580a515ed/97954de1-7b4b-4615-8d0f-0e8580a515ed.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 941.457361] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 941.457618] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c4402b7a-0f20-4d67-9ffc-2eeac5d6dad5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.464397] env[62965]: DEBUG oslo_vmware.api [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 941.464397] env[62965]: value = "task-1390098" [ 941.464397] env[62965]: _type = "Task" [ 941.464397] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.474517] env[62965]: DEBUG oslo_vmware.api [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390098, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.588398] env[62965]: DEBUG nova.compute.manager [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 941.604103] env[62965]: DEBUG oslo_vmware.api [None req-e73a534f-af98-4393-85f0-5e6378f4083e tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390096, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.613337] env[62965]: DEBUG nova.virt.hardware [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 941.613732] env[62965]: DEBUG nova.virt.hardware [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 941.613948] env[62965]: DEBUG nova.virt.hardware [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 941.614284] env[62965]: DEBUG nova.virt.hardware [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 941.614449] env[62965]: DEBUG nova.virt.hardware [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 941.614641] env[62965]: DEBUG nova.virt.hardware [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 941.614897] env[62965]: DEBUG nova.virt.hardware [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 941.615112] env[62965]: DEBUG nova.virt.hardware [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 941.615370] env[62965]: DEBUG nova.virt.hardware [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 941.615576] env[62965]: DEBUG nova.virt.hardware [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 941.615790] env[62965]: DEBUG nova.virt.hardware [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 941.616742] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98775a93-e365-4eb4-bb51-1acf08223e26 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.625273] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5b35046-6f77-4c1e-91ee-728ffb892d8b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.801407] env[62965]: DEBUG nova.compute.manager [req-b86378da-c406-47fb-89ec-9fca5711e345 req-e6ce7837-50a1-48c8-8b1c-7a813f9bc999 service nova] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Received event network-vif-plugged-3eafc53f-b78d-43b3-9341-89b3a9cd0ecb {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 941.801921] env[62965]: DEBUG oslo_concurrency.lockutils [req-b86378da-c406-47fb-89ec-9fca5711e345 req-e6ce7837-50a1-48c8-8b1c-7a813f9bc999 service nova] Acquiring lock "3223b1ed-abc1-439c-bcad-6317eb204e45-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.802219] env[62965]: DEBUG oslo_concurrency.lockutils [req-b86378da-c406-47fb-89ec-9fca5711e345 req-e6ce7837-50a1-48c8-8b1c-7a813f9bc999 service nova] Lock "3223b1ed-abc1-439c-bcad-6317eb204e45-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.802493] env[62965]: DEBUG oslo_concurrency.lockutils [req-b86378da-c406-47fb-89ec-9fca5711e345 req-e6ce7837-50a1-48c8-8b1c-7a813f9bc999 service nova] Lock "3223b1ed-abc1-439c-bcad-6317eb204e45-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.802706] env[62965]: DEBUG nova.compute.manager [req-b86378da-c406-47fb-89ec-9fca5711e345 req-e6ce7837-50a1-48c8-8b1c-7a813f9bc999 service nova] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] No waiting events found dispatching network-vif-plugged-3eafc53f-b78d-43b3-9341-89b3a9cd0ecb {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 941.802937] env[62965]: WARNING nova.compute.manager [req-b86378da-c406-47fb-89ec-9fca5711e345 req-e6ce7837-50a1-48c8-8b1c-7a813f9bc999 service nova] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Received unexpected event network-vif-plugged-3eafc53f-b78d-43b3-9341-89b3a9cd0ecb for instance with vm_state building and task_state spawning. [ 941.872148] env[62965]: DEBUG nova.scheduler.client.report [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 941.889857] env[62965]: DEBUG oslo_concurrency.lockutils [req-4ded6c47-6c91-4cc1-a02c-bd51803361b0 req-00b1e0d7-f043-445b-9b41-d08806ffe78a service nova] Releasing lock "refresh_cache-97954de1-7b4b-4615-8d0f-0e8580a515ed" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.975232] env[62965]: DEBUG oslo_vmware.api [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390098, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.144282} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.975668] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 941.976860] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b955afc0-e78e-4cf1-8819-fccc1d6e715c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.003457] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] 97954de1-7b4b-4615-8d0f-0e8580a515ed/97954de1-7b4b-4615-8d0f-0e8580a515ed.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 942.006497] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3475f72c-8b82-4e40-9bb9-87af361fc0ef {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.032590] env[62965]: DEBUG oslo_vmware.api [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 942.032590] env[62965]: value = "task-1390099" [ 942.032590] env[62965]: _type = "Task" [ 942.032590] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.042868] env[62965]: DEBUG oslo_vmware.api [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390099, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.100334] env[62965]: DEBUG nova.network.neutron [req-b98ff61b-7973-407b-ad0d-7874d8f8f507 req-88d95682-ca7d-4567-858f-bfc35a11cb09 service nova] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Updated VIF entry in instance network info cache for port d2d44211-6856-42ed-8f9f-3b1a58b31349. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 942.100900] env[62965]: DEBUG nova.network.neutron [req-b98ff61b-7973-407b-ad0d-7874d8f8f507 req-88d95682-ca7d-4567-858f-bfc35a11cb09 service nova] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Updating instance_info_cache with network_info: [{"id": "d2d44211-6856-42ed-8f9f-3b1a58b31349", "address": "fa:16:3e:3b:f0:e9", "network": {"id": "49dd42b0-6397-4fdd-a319-ed920c72a5af", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-373172566-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.230", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d91c99b14018469e8998fd1ccc0fab4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapd2d44211-68", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.108098] env[62965]: DEBUG oslo_vmware.api [None req-e73a534f-af98-4393-85f0-5e6378f4083e tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390096, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.377988] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.813s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.380741] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.269s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.381043] env[62965]: DEBUG nova.objects.instance [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lazy-loading 'resources' on Instance uuid b59d6023-a047-4b32-bd18-1c7cc05f67d7 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 942.387914] env[62965]: DEBUG nova.network.neutron [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Successfully updated port: 3eafc53f-b78d-43b3-9341-89b3a9cd0ecb {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 942.410076] env[62965]: INFO nova.scheduler.client.report [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Deleted allocations for instance 64cf898c-2c97-4daa-80c8-0a1c252cbcf2 [ 942.543906] env[62965]: DEBUG oslo_vmware.api [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390099, 'name': ReconfigVM_Task, 'duration_secs': 0.470735} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.544638] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Reconfigured VM instance instance-00000056 to attach disk [datastore1] 97954de1-7b4b-4615-8d0f-0e8580a515ed/97954de1-7b4b-4615-8d0f-0e8580a515ed.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 942.545440] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5f7a4df7-b5ae-431e-8ace-0612478e455f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.554136] env[62965]: DEBUG oslo_vmware.api [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 942.554136] env[62965]: value = "task-1390100" [ 942.554136] env[62965]: _type = "Task" [ 942.554136] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.571407] env[62965]: DEBUG oslo_vmware.api [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390100, 'name': Rename_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.605861] env[62965]: DEBUG oslo_vmware.api [None req-e73a534f-af98-4393-85f0-5e6378f4083e tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390096, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.609709] env[62965]: DEBUG oslo_concurrency.lockutils [req-b98ff61b-7973-407b-ad0d-7874d8f8f507 req-88d95682-ca7d-4567-858f-bfc35a11cb09 service nova] Releasing lock "refresh_cache-b59d6023-a047-4b32-bd18-1c7cc05f67d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.609968] env[62965]: DEBUG nova.compute.manager [req-b98ff61b-7973-407b-ad0d-7874d8f8f507 req-88d95682-ca7d-4567-858f-bfc35a11cb09 service nova] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Received event network-vif-deleted-849e89a5-b45f-482f-a43b-4fd17a659c17 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 942.884641] env[62965]: DEBUG nova.objects.instance [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lazy-loading 'numa_topology' on Instance uuid b59d6023-a047-4b32-bd18-1c7cc05f67d7 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 942.890849] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Acquiring lock "refresh_cache-3223b1ed-abc1-439c-bcad-6317eb204e45" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.890983] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Acquired lock "refresh_cache-3223b1ed-abc1-439c-bcad-6317eb204e45" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.891170] env[62965]: DEBUG nova.network.neutron [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 942.922111] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9ceb723d-d50d-449c-9eff-5df816a5f0bf tempest-AttachVolumeTestJSON-1565951517 tempest-AttachVolumeTestJSON-1565951517-project-member] Lock "64cf898c-2c97-4daa-80c8-0a1c252cbcf2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.104s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.065602] env[62965]: DEBUG oslo_vmware.api [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390100, 'name': Rename_Task, 'duration_secs': 0.305434} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.065969] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 943.066351] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c19ad6dc-eeab-421f-b0a7-3488cc62483f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.074929] env[62965]: DEBUG oslo_vmware.api [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 943.074929] env[62965]: value = "task-1390101" [ 943.074929] env[62965]: _type = "Task" [ 943.074929] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.083762] env[62965]: DEBUG oslo_vmware.api [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390101, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.085393] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a4245b41-dd2b-44aa-b46d-f458a24a3a34 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "fbbcbe60-92f3-497d-b190-520a2547b3a9" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.085617] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a4245b41-dd2b-44aa-b46d-f458a24a3a34 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "fbbcbe60-92f3-497d-b190-520a2547b3a9" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.085802] env[62965]: DEBUG nova.compute.manager [None req-a4245b41-dd2b-44aa-b46d-f458a24a3a34 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Going to confirm migration 1 {{(pid=62965) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5112}} [ 943.105189] env[62965]: DEBUG oslo_vmware.api [None req-e73a534f-af98-4393-85f0-5e6378f4083e tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390096, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.326623] env[62965]: DEBUG nova.compute.manager [req-1532b27d-0aac-45c1-9ff5-05f95e445d38 req-9a33e938-be5c-4cb9-be12-b8a276b41e53 service nova] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Received event network-changed-3eafc53f-b78d-43b3-9341-89b3a9cd0ecb {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 943.326864] env[62965]: DEBUG nova.compute.manager [req-1532b27d-0aac-45c1-9ff5-05f95e445d38 req-9a33e938-be5c-4cb9-be12-b8a276b41e53 service nova] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Refreshing instance network info cache due to event network-changed-3eafc53f-b78d-43b3-9341-89b3a9cd0ecb. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 943.327103] env[62965]: DEBUG oslo_concurrency.lockutils [req-1532b27d-0aac-45c1-9ff5-05f95e445d38 req-9a33e938-be5c-4cb9-be12-b8a276b41e53 service nova] Acquiring lock "refresh_cache-3223b1ed-abc1-439c-bcad-6317eb204e45" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.358147] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-20a96db8-5668-4ec2-8934-9f36f0fb13e0 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Volume attach. Driver type: vmdk {{(pid=62965) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 943.358555] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-20a96db8-5668-4ec2-8934-9f36f0fb13e0 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-295022', 'volume_id': 'faaca442-1091-4470-8417-2b2608e503c8', 'name': 'volume-faaca442-1091-4470-8417-2b2608e503c8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0eb943ae-462c-40dd-a870-7a884ea2db74', 'attached_at': '', 'detached_at': '', 'volume_id': 'faaca442-1091-4470-8417-2b2608e503c8', 'serial': 'faaca442-1091-4470-8417-2b2608e503c8'} {{(pid=62965) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 943.359861] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82b9f117-e4e6-445e-bc5d-938f81d68b7f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.378552] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17948f85-9d45-46bd-8366-d4283aa89175 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.398536] env[62965]: DEBUG nova.objects.base [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=62965) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 943.410210] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-20a96db8-5668-4ec2-8934-9f36f0fb13e0 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] volume-faaca442-1091-4470-8417-2b2608e503c8/volume-faaca442-1091-4470-8417-2b2608e503c8.vmdk or device None with type thin {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 943.410826] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c96e472c-6bf0-4c02-8033-f1befca741ad {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.432186] env[62965]: DEBUG oslo_vmware.api [None req-20a96db8-5668-4ec2-8934-9f36f0fb13e0 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 943.432186] env[62965]: value = "task-1390102" [ 943.432186] env[62965]: _type = "Task" [ 943.432186] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.442282] env[62965]: DEBUG oslo_vmware.api [None req-20a96db8-5668-4ec2-8934-9f36f0fb13e0 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390102, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.466166] env[62965]: DEBUG nova.network.neutron [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 943.591291] env[62965]: DEBUG oslo_vmware.api [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390101, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.609672] env[62965]: DEBUG oslo_vmware.api [None req-e73a534f-af98-4393-85f0-5e6378f4083e tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390096, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.640486] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75764677-e0cc-4bf7-8f64-06b1cfcc6ce3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.650913] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e17dbdd4-5361-44d6-9a1b-5ae4b14cbeff {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.688521] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-459c135d-7cf3-4022-962e-4713e40940cd {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.697715] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2282855-c6d1-4add-a1d9-0581a2a5d2f8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.704932] env[62965]: DEBUG nova.network.neutron [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Updating instance_info_cache with network_info: [{"id": "3eafc53f-b78d-43b3-9341-89b3a9cd0ecb", "address": "fa:16:3e:c9:a2:be", "network": {"id": "a27e0199-89db-4b64-b3fd-4a12cd598e60", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-12171013-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d3edf0643c9d40e9bb568742b7a7a508", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f4a795c-8718-4a7c-aafe-9da231df10f8", "external-id": "nsx-vlan-transportzone-162", "segmentation_id": 162, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3eafc53f-b7", "ovs_interfaceid": "3eafc53f-b78d-43b3-9341-89b3a9cd0ecb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.720775] env[62965]: DEBUG nova.compute.provider_tree [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 943.720852] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a4245b41-dd2b-44aa-b46d-f458a24a3a34 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "refresh_cache-fbbcbe60-92f3-497d-b190-520a2547b3a9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.721298] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a4245b41-dd2b-44aa-b46d-f458a24a3a34 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquired lock "refresh_cache-fbbcbe60-92f3-497d-b190-520a2547b3a9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.721298] env[62965]: DEBUG nova.network.neutron [None req-a4245b41-dd2b-44aa-b46d-f458a24a3a34 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 943.721498] env[62965]: DEBUG nova.objects.instance [None req-a4245b41-dd2b-44aa-b46d-f458a24a3a34 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lazy-loading 'info_cache' on Instance uuid fbbcbe60-92f3-497d-b190-520a2547b3a9 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 943.944115] env[62965]: DEBUG oslo_vmware.api [None req-20a96db8-5668-4ec2-8934-9f36f0fb13e0 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390102, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.086819] env[62965]: DEBUG oslo_vmware.api [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390101, 'name': PowerOnVM_Task, 'duration_secs': 0.761215} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.087214] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 944.087474] env[62965]: INFO nova.compute.manager [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Took 8.46 seconds to spawn the instance on the hypervisor. [ 944.087703] env[62965]: DEBUG nova.compute.manager [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 944.088515] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b637266c-5f90-4925-aeaf-76dbf78ff174 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.106868] env[62965]: DEBUG oslo_vmware.api [None req-e73a534f-af98-4393-85f0-5e6378f4083e tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390096, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.208525] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Releasing lock "refresh_cache-3223b1ed-abc1-439c-bcad-6317eb204e45" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 944.208822] env[62965]: DEBUG nova.compute.manager [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Instance network_info: |[{"id": "3eafc53f-b78d-43b3-9341-89b3a9cd0ecb", "address": "fa:16:3e:c9:a2:be", "network": {"id": "a27e0199-89db-4b64-b3fd-4a12cd598e60", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-12171013-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d3edf0643c9d40e9bb568742b7a7a508", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f4a795c-8718-4a7c-aafe-9da231df10f8", "external-id": "nsx-vlan-transportzone-162", "segmentation_id": 162, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3eafc53f-b7", "ovs_interfaceid": "3eafc53f-b78d-43b3-9341-89b3a9cd0ecb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 944.209158] env[62965]: DEBUG oslo_concurrency.lockutils [req-1532b27d-0aac-45c1-9ff5-05f95e445d38 req-9a33e938-be5c-4cb9-be12-b8a276b41e53 service nova] Acquired lock "refresh_cache-3223b1ed-abc1-439c-bcad-6317eb204e45" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.209355] env[62965]: DEBUG nova.network.neutron [req-1532b27d-0aac-45c1-9ff5-05f95e445d38 req-9a33e938-be5c-4cb9-be12-b8a276b41e53 service nova] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Refreshing network info cache for port 3eafc53f-b78d-43b3-9341-89b3a9cd0ecb {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 944.210624] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c9:a2:be', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3f4a795c-8718-4a7c-aafe-9da231df10f8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3eafc53f-b78d-43b3-9341-89b3a9cd0ecb', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 944.219102] env[62965]: DEBUG oslo.service.loopingcall [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 944.220201] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 944.220479] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-727fbcc7-3f37-4b0a-80bc-e7d65a1c2465 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.235802] env[62965]: DEBUG nova.scheduler.client.report [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 944.248513] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 944.248513] env[62965]: value = "task-1390104" [ 944.248513] env[62965]: _type = "Task" [ 944.248513] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.259153] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390104, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.443926] env[62965]: DEBUG oslo_vmware.api [None req-20a96db8-5668-4ec2-8934-9f36f0fb13e0 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390102, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.609718] env[62965]: INFO nova.compute.manager [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Took 13.41 seconds to build instance. [ 944.616365] env[62965]: DEBUG oslo_vmware.api [None req-e73a534f-af98-4393-85f0-5e6378f4083e tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390096, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.745686] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.365s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.748969] env[62965]: DEBUG oslo_concurrency.lockutils [None req-0ed4cac0-5b86-449d-b75b-6ea9aaa3765c tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.798s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.748969] env[62965]: DEBUG nova.objects.instance [None req-0ed4cac0-5b86-449d-b75b-6ea9aaa3765c tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lazy-loading 'resources' on Instance uuid 1be808db-d619-4c56-8092-ca52809ec10e {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 944.764813] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390104, 'name': CreateVM_Task, 'duration_secs': 0.369346} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.764813] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 944.767984] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.767984] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.767984] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 944.767984] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eabccb9d-357b-45fb-98b4-b1deac69874e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.771906] env[62965]: DEBUG oslo_vmware.api [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 944.771906] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5283e270-e018-96f6-f6c4-b22bde020106" [ 944.771906] env[62965]: _type = "Task" [ 944.771906] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.783386] env[62965]: DEBUG oslo_vmware.api [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5283e270-e018-96f6-f6c4-b22bde020106, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.947495] env[62965]: DEBUG oslo_vmware.api [None req-20a96db8-5668-4ec2-8934-9f36f0fb13e0 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390102, 'name': ReconfigVM_Task, 'duration_secs': 1.463837} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.947906] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-20a96db8-5668-4ec2-8934-9f36f0fb13e0 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Reconfigured VM instance instance-00000054 to attach disk [datastore1] volume-faaca442-1091-4470-8417-2b2608e503c8/volume-faaca442-1091-4470-8417-2b2608e503c8.vmdk or device None with type thin {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 944.955661] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-072fe94c-5c61-40b9-95bf-bdc0aff5579d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.972171] env[62965]: DEBUG oslo_vmware.api [None req-20a96db8-5668-4ec2-8934-9f36f0fb13e0 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 944.972171] env[62965]: value = "task-1390105" [ 944.972171] env[62965]: _type = "Task" [ 944.972171] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.981086] env[62965]: DEBUG oslo_vmware.api [None req-20a96db8-5668-4ec2-8934-9f36f0fb13e0 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390105, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.109921] env[62965]: DEBUG nova.network.neutron [req-1532b27d-0aac-45c1-9ff5-05f95e445d38 req-9a33e938-be5c-4cb9-be12-b8a276b41e53 service nova] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Updated VIF entry in instance network info cache for port 3eafc53f-b78d-43b3-9341-89b3a9cd0ecb. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 945.110497] env[62965]: DEBUG nova.network.neutron [req-1532b27d-0aac-45c1-9ff5-05f95e445d38 req-9a33e938-be5c-4cb9-be12-b8a276b41e53 service nova] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Updating instance_info_cache with network_info: [{"id": "3eafc53f-b78d-43b3-9341-89b3a9cd0ecb", "address": "fa:16:3e:c9:a2:be", "network": {"id": "a27e0199-89db-4b64-b3fd-4a12cd598e60", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-12171013-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d3edf0643c9d40e9bb568742b7a7a508", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3f4a795c-8718-4a7c-aafe-9da231df10f8", "external-id": "nsx-vlan-transportzone-162", "segmentation_id": 162, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3eafc53f-b7", "ovs_interfaceid": "3eafc53f-b78d-43b3-9341-89b3a9cd0ecb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.112134] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b55f73e8-f241-40d9-acb9-3d401393a7ee tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "97954de1-7b4b-4615-8d0f-0e8580a515ed" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.918s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.116534] env[62965]: DEBUG oslo_vmware.api [None req-e73a534f-af98-4393-85f0-5e6378f4083e tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390096, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.117416] env[62965]: DEBUG nova.network.neutron [None req-a4245b41-dd2b-44aa-b46d-f458a24a3a34 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Updating instance_info_cache with network_info: [{"id": "3edb59ed-a2b8-4d31-b975-9c9d64c3a22b", "address": "fa:16:3e:92:26:4b", "network": {"id": "a5246cc1-6001-4c20-9724-fd36a0437582", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-644702631-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bfa5cfdc7d5496482d87675f2a122dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3edb59ed-a2", "ovs_interfaceid": "3edb59ed-a2b8-4d31-b975-9c9d64c3a22b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.257931] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b1cbda11-06a3-4446-8c9c-7c06d3e3390d tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "b59d6023-a047-4b32-bd18-1c7cc05f67d7" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 24.946s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.258775] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "b59d6023-a047-4b32-bd18-1c7cc05f67d7" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 4.109s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.258974] env[62965]: INFO nova.compute.manager [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Unshelving [ 945.285249] env[62965]: DEBUG oslo_vmware.api [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5283e270-e018-96f6-f6c4-b22bde020106, 'name': SearchDatastore_Task, 'duration_secs': 0.0118} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.286032] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.286032] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 945.286188] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.286229] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.286733] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 945.286733] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-36a16643-996e-40e9-9c50-3b8e7c52af69 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.296594] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 945.297156] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 945.297563] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a79a1b7-7a2d-4867-93c6-c9a704a3cdc3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.303260] env[62965]: DEBUG oslo_vmware.api [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 945.303260] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52647b5d-7d5b-0a68-9684-653421450959" [ 945.303260] env[62965]: _type = "Task" [ 945.303260] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.314594] env[62965]: DEBUG oslo_vmware.api [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52647b5d-7d5b-0a68-9684-653421450959, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.440563] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f7355f6-4f67-460d-9dac-85291009a0eb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.448549] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfa1fbb7-04dc-4512-acdf-6950a7c2cdd9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.483236] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac00d9e0-f4a3-4421-be59-7e230f53d289 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.493739] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-738add95-8fcb-440d-85b9-0c6638756186 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.497441] env[62965]: DEBUG oslo_vmware.api [None req-20a96db8-5668-4ec2-8934-9f36f0fb13e0 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390105, 'name': ReconfigVM_Task, 'duration_secs': 0.160779} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.497737] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-20a96db8-5668-4ec2-8934-9f36f0fb13e0 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-295022', 'volume_id': 'faaca442-1091-4470-8417-2b2608e503c8', 'name': 'volume-faaca442-1091-4470-8417-2b2608e503c8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0eb943ae-462c-40dd-a870-7a884ea2db74', 'attached_at': '', 'detached_at': '', 'volume_id': 'faaca442-1091-4470-8417-2b2608e503c8', 'serial': 'faaca442-1091-4470-8417-2b2608e503c8'} {{(pid=62965) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 945.509124] env[62965]: DEBUG nova.compute.provider_tree [None req-0ed4cac0-5b86-449d-b75b-6ea9aaa3765c tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 945.616250] env[62965]: DEBUG oslo_concurrency.lockutils [req-1532b27d-0aac-45c1-9ff5-05f95e445d38 req-9a33e938-be5c-4cb9-be12-b8a276b41e53 service nova] Releasing lock "refresh_cache-3223b1ed-abc1-439c-bcad-6317eb204e45" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.616935] env[62965]: DEBUG oslo_vmware.api [None req-e73a534f-af98-4393-85f0-5e6378f4083e tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390096, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.620321] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a4245b41-dd2b-44aa-b46d-f458a24a3a34 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Releasing lock "refresh_cache-fbbcbe60-92f3-497d-b190-520a2547b3a9" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.620551] env[62965]: DEBUG nova.objects.instance [None req-a4245b41-dd2b-44aa-b46d-f458a24a3a34 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lazy-loading 'migration_context' on Instance uuid fbbcbe60-92f3-497d-b190-520a2547b3a9 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 945.814291] env[62965]: DEBUG oslo_vmware.api [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52647b5d-7d5b-0a68-9684-653421450959, 'name': SearchDatastore_Task, 'duration_secs': 0.010397} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.815191] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2fb7f17a-68a5-458b-8df6-71e93c661d40 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.821026] env[62965]: DEBUG oslo_vmware.api [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 945.821026] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]522d22ef-046f-65eb-b75b-62ec5baf6194" [ 945.821026] env[62965]: _type = "Task" [ 945.821026] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.829200] env[62965]: DEBUG oslo_vmware.api [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]522d22ef-046f-65eb-b75b-62ec5baf6194, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.011727] env[62965]: DEBUG nova.scheduler.client.report [None req-0ed4cac0-5b86-449d-b75b-6ea9aaa3765c tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 946.061233] env[62965]: DEBUG nova.compute.manager [None req-f99d4b6d-896b-445f-bf43-4d0b2692cbd7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 946.062365] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f1a6876-90b5-47e5-93e3-e266b848eede {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.118558] env[62965]: DEBUG oslo_vmware.api [None req-e73a534f-af98-4393-85f0-5e6378f4083e tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390096, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.123625] env[62965]: DEBUG nova.objects.base [None req-a4245b41-dd2b-44aa-b46d-f458a24a3a34 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62965) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 946.124523] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10f92966-ffa1-4303-8f62-2f47c740a95a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.146241] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3402181b-01ce-4d75-84a6-1a3cf54a90b1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.154204] env[62965]: DEBUG oslo_vmware.api [None req-a4245b41-dd2b-44aa-b46d-f458a24a3a34 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 946.154204] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52e8a57f-5e6b-c0b8-75fe-1715da0b5545" [ 946.154204] env[62965]: _type = "Task" [ 946.154204] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.165695] env[62965]: DEBUG oslo_vmware.api [None req-a4245b41-dd2b-44aa-b46d-f458a24a3a34 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52e8a57f-5e6b-c0b8-75fe-1715da0b5545, 'name': SearchDatastore_Task, 'duration_secs': 0.007462} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.166050] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a4245b41-dd2b-44aa-b46d-f458a24a3a34 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.270408] env[62965]: DEBUG nova.compute.utils [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 946.333325] env[62965]: DEBUG oslo_vmware.api [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]522d22ef-046f-65eb-b75b-62ec5baf6194, 'name': SearchDatastore_Task, 'duration_secs': 0.009934} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.333550] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.333740] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 3223b1ed-abc1-439c-bcad-6317eb204e45/3223b1ed-abc1-439c-bcad-6317eb204e45.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 946.334059] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cae9c2a9-f0b0-4252-a2bb-8ab220acda54 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.342991] env[62965]: DEBUG oslo_vmware.api [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 946.342991] env[62965]: value = "task-1390106" [ 946.342991] env[62965]: _type = "Task" [ 946.342991] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.352765] env[62965]: DEBUG oslo_vmware.api [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390106, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.517265] env[62965]: DEBUG oslo_concurrency.lockutils [None req-0ed4cac0-5b86-449d-b75b-6ea9aaa3765c tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.768s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.520387] env[62965]: DEBUG oslo_concurrency.lockutils [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.807s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.523049] env[62965]: INFO nova.compute.claims [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 946.536410] env[62965]: DEBUG nova.objects.instance [None req-20a96db8-5668-4ec2-8934-9f36f0fb13e0 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lazy-loading 'flavor' on Instance uuid 0eb943ae-462c-40dd-a870-7a884ea2db74 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 946.545122] env[62965]: INFO nova.scheduler.client.report [None req-0ed4cac0-5b86-449d-b75b-6ea9aaa3765c tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Deleted allocations for instance 1be808db-d619-4c56-8092-ca52809ec10e [ 946.574470] env[62965]: INFO nova.compute.manager [None req-f99d4b6d-896b-445f-bf43-4d0b2692cbd7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] instance snapshotting [ 946.577619] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19854519-9aee-4881-80dd-bb9eb99057d2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.600796] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b00adf0-87c8-484f-a048-b8e27008b72c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.622482] env[62965]: DEBUG oslo_vmware.api [None req-e73a534f-af98-4393-85f0-5e6378f4083e tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390096, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.775035] env[62965]: INFO nova.virt.block_device [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Booting with volume 437ab7f8-7ad9-4d87-8f07-900702627943 at /dev/sdb [ 946.814199] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d5117999-8843-4433-827e-de5318bc2fac {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.826043] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94f9ed1d-5f37-49f3-8f5e-37083c5c3163 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.853158] env[62965]: DEBUG oslo_vmware.api [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390106, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.486224} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.864368] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 3223b1ed-abc1-439c-bcad-6317eb204e45/3223b1ed-abc1-439c-bcad-6317eb204e45.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 946.864650] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 946.865126] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-257ac2b4-abda-4517-a0c5-630fa5e8b004 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.867129] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-94c51487-ab81-4131-b0fc-335d22401342 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.877214] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f68c2b3b-c79d-4d35-8328-3f9634a3f82c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.889908] env[62965]: DEBUG oslo_vmware.api [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 946.889908] env[62965]: value = "task-1390107" [ 946.889908] env[62965]: _type = "Task" [ 946.889908] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.899876] env[62965]: DEBUG oslo_vmware.api [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390107, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.914535] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 946.915172] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 946.915380] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Starting heal instance info cache {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10257}} [ 946.915497] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Rebuilding the list of instances to heal {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10261}} [ 946.918195] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcede89a-da0f-4856-981d-fb408aad3113 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.927246] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef0f53c0-11f8-4da0-bea0-1c1a3d5801d4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.943820] env[62965]: DEBUG nova.virt.block_device [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Updating existing volume attachment record: d538e9d3-8c0c-435b-89fa-8d324bd35d79 {{(pid=62965) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 947.042166] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20a96db8-5668-4ec2-8934-9f36f0fb13e0 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "0eb943ae-462c-40dd-a870-7a884ea2db74" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.299s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.052588] env[62965]: DEBUG oslo_concurrency.lockutils [None req-0ed4cac0-5b86-449d-b75b-6ea9aaa3765c tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "1be808db-d619-4c56-8092-ca52809ec10e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.556s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.117798] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-f99d4b6d-896b-445f-bf43-4d0b2692cbd7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Creating Snapshot of the VM instance {{(pid=62965) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 947.118182] env[62965]: DEBUG oslo_vmware.api [None req-e73a534f-af98-4393-85f0-5e6378f4083e tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390096, 'name': ReconfigVM_Task, 'duration_secs': 5.902946} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.118439] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-1e176020-6187-44c9-867e-ae3e9fa56362 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.120966] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e73a534f-af98-4393-85f0-5e6378f4083e tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Releasing lock "d038608c-9264-4fd5-b9ba-0653c98ec0e4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.121107] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e73a534f-af98-4393-85f0-5e6378f4083e tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Reconfigured VM to detach interface {{(pid=62965) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 947.128757] env[62965]: DEBUG oslo_vmware.api [None req-f99d4b6d-896b-445f-bf43-4d0b2692cbd7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 947.128757] env[62965]: value = "task-1390110" [ 947.128757] env[62965]: _type = "Task" [ 947.128757] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.139205] env[62965]: DEBUG oslo_vmware.api [None req-f99d4b6d-896b-445f-bf43-4d0b2692cbd7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390110, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.205413] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "0eb943ae-462c-40dd-a870-7a884ea2db74" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.205711] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "0eb943ae-462c-40dd-a870-7a884ea2db74" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.205932] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "0eb943ae-462c-40dd-a870-7a884ea2db74-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.206134] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "0eb943ae-462c-40dd-a870-7a884ea2db74-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.206312] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "0eb943ae-462c-40dd-a870-7a884ea2db74-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.212694] env[62965]: INFO nova.compute.manager [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Terminating instance [ 947.404047] env[62965]: DEBUG oslo_vmware.api [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390107, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.095049} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.404637] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 947.405530] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-213cd35d-5178-453b-afa9-2bc0574d1b3f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.425560] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Skipping network cache update for instance because it is being deleted. {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10274}} [ 947.425797] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Skipping network cache update for instance because it is Building. {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10270}} [ 947.426250] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Skipping network cache update for instance because it is Building. {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10270}} [ 947.438597] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] 3223b1ed-abc1-439c-bcad-6317eb204e45/3223b1ed-abc1-439c-bcad-6317eb204e45.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 947.438597] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0208c027-b1e5-43fc-8e13-e984959bd51a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.463323] env[62965]: DEBUG oslo_vmware.api [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 947.463323] env[62965]: value = "task-1390113" [ 947.463323] env[62965]: _type = "Task" [ 947.463323] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.473272] env[62965]: DEBUG oslo_vmware.api [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390113, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.480549] env[62965]: DEBUG oslo_concurrency.lockutils [None req-085cd9fc-7ae6-490b-8cdf-8365bf1ee3c0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "c2448b76-1553-4b68-a731-0bb0967f4c1d" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.480549] env[62965]: DEBUG oslo_concurrency.lockutils [None req-085cd9fc-7ae6-490b-8cdf-8365bf1ee3c0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "c2448b76-1553-4b68-a731-0bb0967f4c1d" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.480549] env[62965]: INFO nova.compute.manager [None req-085cd9fc-7ae6-490b-8cdf-8365bf1ee3c0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Rebooting instance [ 947.501171] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Acquiring lock "refresh_cache-c2448b76-1553-4b68-a731-0bb0967f4c1d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.501421] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Acquired lock "refresh_cache-c2448b76-1553-4b68-a731-0bb0967f4c1d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.501665] env[62965]: DEBUG nova.network.neutron [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Forcefully refreshing network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 947.501918] env[62965]: DEBUG nova.objects.instance [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lazy-loading 'info_cache' on Instance uuid c2448b76-1553-4b68-a731-0bb0967f4c1d {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 947.638915] env[62965]: DEBUG oslo_vmware.api [None req-f99d4b6d-896b-445f-bf43-4d0b2692cbd7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390110, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.718276] env[62965]: DEBUG nova.compute.manager [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 947.718647] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 947.718943] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6fa29567-9649-4be6-9e0a-f63dae03df66 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.727718] env[62965]: DEBUG oslo_vmware.api [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 947.727718] env[62965]: value = "task-1390114" [ 947.727718] env[62965]: _type = "Task" [ 947.727718] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.735099] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efba8676-a5a7-4f4e-8413-408937064c67 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.740762] env[62965]: DEBUG oslo_vmware.api [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390114, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.746488] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37c97b2b-8dac-480f-934c-b669f98eae87 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.783469] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5507eca-5206-4ad4-ad70-4f573ce360c3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.792859] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-912b8ac2-7bd2-4b88-9c33-9bb9efea110d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.808569] env[62965]: DEBUG nova.compute.provider_tree [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 947.848680] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "08f7b05f-3fb5-48ac-a68a-1d6fd804a622" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.849042] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "08f7b05f-3fb5-48ac-a68a-1d6fd804a622" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.974813] env[62965]: DEBUG oslo_vmware.api [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390113, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.006062] env[62965]: DEBUG oslo_concurrency.lockutils [None req-085cd9fc-7ae6-490b-8cdf-8365bf1ee3c0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "refresh_cache-c2448b76-1553-4b68-a731-0bb0967f4c1d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.140439] env[62965]: DEBUG oslo_vmware.api [None req-f99d4b6d-896b-445f-bf43-4d0b2692cbd7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390110, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.243239] env[62965]: DEBUG oslo_vmware.api [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390114, 'name': PowerOffVM_Task, 'duration_secs': 0.265144} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.243523] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 948.243720] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Volume detach. Driver type: vmdk {{(pid=62965) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 948.243912] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-295022', 'volume_id': 'faaca442-1091-4470-8417-2b2608e503c8', 'name': 'volume-faaca442-1091-4470-8417-2b2608e503c8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0eb943ae-462c-40dd-a870-7a884ea2db74', 'attached_at': '', 'detached_at': '', 'volume_id': 'faaca442-1091-4470-8417-2b2608e503c8', 'serial': 'faaca442-1091-4470-8417-2b2608e503c8'} {{(pid=62965) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 948.244711] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd4a09ca-db05-474b-8de6-611fa7a50cd0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.266182] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7d3377d-59c1-411c-ab11-4d37864e9046 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.275512] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-237cadc3-c9b6-4ced-92c2-cc88f437a473 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.298817] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a16434f-a267-4a5c-8f11-feec64642a67 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.316359] env[62965]: DEBUG nova.scheduler.client.report [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 948.319787] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] The volume has not been displaced from its original location: [datastore1] volume-faaca442-1091-4470-8417-2b2608e503c8/volume-faaca442-1091-4470-8417-2b2608e503c8.vmdk. No consolidation needed. {{(pid=62965) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 948.325269] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Reconfiguring VM instance instance-00000054 to detach disk 2001 {{(pid=62965) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 948.326017] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-985ccc10-38d8-4820-a61f-685e9159cdeb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.346736] env[62965]: DEBUG oslo_vmware.api [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 948.346736] env[62965]: value = "task-1390115" [ 948.346736] env[62965]: _type = "Task" [ 948.346736] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.351056] env[62965]: DEBUG nova.compute.manager [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 948.359795] env[62965]: DEBUG oslo_vmware.api [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390115, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.475528] env[62965]: DEBUG oslo_vmware.api [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390113, 'name': ReconfigVM_Task, 'duration_secs': 0.591234} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.476021] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Reconfigured VM instance instance-00000057 to attach disk [datastore2] 3223b1ed-abc1-439c-bcad-6317eb204e45/3223b1ed-abc1-439c-bcad-6317eb204e45.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 948.476604] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7a2246ba-1267-4de4-9235-689f457a97f6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.486489] env[62965]: DEBUG oslo_vmware.api [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 948.486489] env[62965]: value = "task-1390116" [ 948.486489] env[62965]: _type = "Task" [ 948.486489] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.496184] env[62965]: DEBUG oslo_vmware.api [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390116, 'name': Rename_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.569203] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e73a534f-af98-4393-85f0-5e6378f4083e tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "refresh_cache-d038608c-9264-4fd5-b9ba-0653c98ec0e4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.569374] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e73a534f-af98-4393-85f0-5e6378f4083e tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquired lock "refresh_cache-d038608c-9264-4fd5-b9ba-0653c98ec0e4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.569557] env[62965]: DEBUG nova.network.neutron [None req-e73a534f-af98-4393-85f0-5e6378f4083e tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 948.641015] env[62965]: DEBUG oslo_vmware.api [None req-f99d4b6d-896b-445f-bf43-4d0b2692cbd7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390110, 'name': CreateSnapshot_Task, 'duration_secs': 1.189967} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.641015] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-f99d4b6d-896b-445f-bf43-4d0b2692cbd7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Created Snapshot of the VM instance {{(pid=62965) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 948.641651] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4992c324-953e-47df-a4b9-96bbe239c77a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.827767] env[62965]: DEBUG oslo_concurrency.lockutils [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.307s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.827888] env[62965]: DEBUG nova.compute.manager [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 948.830820] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a4245b41-dd2b-44aa-b46d-f458a24a3a34 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 2.665s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.859719] env[62965]: DEBUG oslo_vmware.api [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390115, 'name': ReconfigVM_Task, 'duration_secs': 0.218644} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.862206] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Reconfigured VM instance instance-00000054 to detach disk 2001 {{(pid=62965) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 948.867386] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fdbee838-5dd1-4506-970e-c514735f49e2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.885318] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.886894] env[62965]: DEBUG oslo_vmware.api [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 948.886894] env[62965]: value = "task-1390117" [ 948.886894] env[62965]: _type = "Task" [ 948.886894] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.897975] env[62965]: DEBUG oslo_vmware.api [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390117, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.997715] env[62965]: DEBUG oslo_vmware.api [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390116, 'name': Rename_Task, 'duration_secs': 0.160029} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.997987] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 948.998422] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4ea6f091-6acd-40c8-856e-a304556b82eb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.007276] env[62965]: DEBUG oslo_vmware.api [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 949.007276] env[62965]: value = "task-1390118" [ 949.007276] env[62965]: _type = "Task" [ 949.007276] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.016268] env[62965]: DEBUG oslo_vmware.api [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390118, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.161048] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-f99d4b6d-896b-445f-bf43-4d0b2692cbd7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Creating linked-clone VM from snapshot {{(pid=62965) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 949.161430] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-a80c4e01-3ac6-4d32-84ff-350e457a4ea0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.171799] env[62965]: DEBUG oslo_vmware.api [None req-f99d4b6d-896b-445f-bf43-4d0b2692cbd7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 949.171799] env[62965]: value = "task-1390119" [ 949.171799] env[62965]: _type = "Task" [ 949.171799] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.188995] env[62965]: DEBUG oslo_vmware.api [None req-f99d4b6d-896b-445f-bf43-4d0b2692cbd7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390119, 'name': CloneVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.335051] env[62965]: DEBUG nova.compute.utils [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 949.342753] env[62965]: DEBUG nova.compute.manager [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 949.343863] env[62965]: DEBUG nova.network.neutron [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 949.406507] env[62965]: DEBUG oslo_vmware.api [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390117, 'name': ReconfigVM_Task, 'duration_secs': 0.316156} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.407312] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-295022', 'volume_id': 'faaca442-1091-4470-8417-2b2608e503c8', 'name': 'volume-faaca442-1091-4470-8417-2b2608e503c8', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0eb943ae-462c-40dd-a870-7a884ea2db74', 'attached_at': '', 'detached_at': '', 'volume_id': 'faaca442-1091-4470-8417-2b2608e503c8', 'serial': 'faaca442-1091-4470-8417-2b2608e503c8'} {{(pid=62965) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 949.407961] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 949.409166] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d27ba67-a216-4fc2-95e0-cab0d9757b2c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.418950] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 949.419829] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bcc0c438-e055-4245-9ea8-4c9fe693e3ce {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.506194] env[62965]: DEBUG nova.policy [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a5e5c6c3198946658126b0024ba82583', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9a539017d4b84f39a71b6f4098395fbe', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 949.517073] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 949.517580] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Deleting contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 949.517918] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Deleting the datastore file [datastore2] 0eb943ae-462c-40dd-a870-7a884ea2db74 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 949.519121] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-421c7889-fe16-4e81-9083-84a62e0870f2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.526970] env[62965]: DEBUG oslo_vmware.api [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390118, 'name': PowerOnVM_Task, 'duration_secs': 0.477731} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.526970] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 949.526970] env[62965]: INFO nova.compute.manager [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Took 7.94 seconds to spawn the instance on the hypervisor. [ 949.526970] env[62965]: DEBUG nova.compute.manager [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 949.528190] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-709a5dce-af3d-4367-a09b-7911bd12de65 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.533295] env[62965]: DEBUG oslo_vmware.api [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 949.533295] env[62965]: value = "task-1390122" [ 949.533295] env[62965]: _type = "Task" [ 949.533295] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.558385] env[62965]: DEBUG oslo_vmware.api [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390122, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.580643] env[62965]: INFO nova.network.neutron [None req-e73a534f-af98-4393-85f0-5e6378f4083e tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Port 9c46a8bf-6b23-45d3-95a4-f0c751895c71 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 949.581331] env[62965]: DEBUG nova.network.neutron [None req-e73a534f-af98-4393-85f0-5e6378f4083e tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Updating instance_info_cache with network_info: [{"id": "b1718061-edf1-4077-b57f-0e54aea365a6", "address": "fa:16:3e:1f:27:16", "network": {"id": "a9699d8b-5e76-4666-a236-33ab8f22345e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1329540850-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1952dedf340f4b7ba0d57af6bba9a749", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a316376e-2ef0-4b1e-b40c-10321ebd7e1a", "external-id": "nsx-vlan-transportzone-942", "segmentation_id": 942, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1718061-ed", "ovs_interfaceid": "b1718061-edf1-4077-b57f-0e54aea365a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.586866] env[62965]: DEBUG nova.network.neutron [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Updating instance_info_cache with network_info: [{"id": "57672e8b-1df1-463d-b9b3-8290667d9fd2", "address": "fa:16:3e:53:46:47", "network": {"id": "4649fed7-c2ca-43a9-a183-7c99423da726", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1978698173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fab418e2fe1420793517663574b43bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43ad01d2-c7dd-453c-a929-8ad76294d13c", "external-id": "nsx-vlan-transportzone-176", "segmentation_id": 176, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57672e8b-1d", "ovs_interfaceid": "57672e8b-1df1-463d-b9b3-8290667d9fd2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.627721] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75602f8b-d7ae-4bf6-b8bb-5330057f3d12 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.637787] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f88d526-c1d4-4f29-bcf4-9a67eb73588c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.681901] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1e34337-771d-4ed3-b26d-4a40917fba52 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.691430] env[62965]: DEBUG oslo_vmware.api [None req-f99d4b6d-896b-445f-bf43-4d0b2692cbd7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390119, 'name': CloneVM_Task} progress is 94%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.694981] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e4b3e4b-71a0-43da-9121-86dda222571d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.711660] env[62965]: DEBUG nova.compute.provider_tree [None req-a4245b41-dd2b-44aa-b46d-f458a24a3a34 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 949.843233] env[62965]: DEBUG nova.compute.manager [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 949.925735] env[62965]: DEBUG nova.network.neutron [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Successfully created port: 219e79ce-5ab0-49eb-a38f-6b1ba814cdf3 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 950.001888] env[62965]: DEBUG nova.compute.manager [req-733fc234-16b9-4330-b6b5-186b70729774 req-a9f22bb2-2f8d-4482-9c5b-cfbba395ea8e service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Received event network-changed-b1718061-edf1-4077-b57f-0e54aea365a6 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 950.001888] env[62965]: DEBUG nova.compute.manager [req-733fc234-16b9-4330-b6b5-186b70729774 req-a9f22bb2-2f8d-4482-9c5b-cfbba395ea8e service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Refreshing instance network info cache due to event network-changed-b1718061-edf1-4077-b57f-0e54aea365a6. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 950.002174] env[62965]: DEBUG oslo_concurrency.lockutils [req-733fc234-16b9-4330-b6b5-186b70729774 req-a9f22bb2-2f8d-4482-9c5b-cfbba395ea8e service nova] Acquiring lock "refresh_cache-d038608c-9264-4fd5-b9ba-0653c98ec0e4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.059019] env[62965]: DEBUG oslo_vmware.api [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390122, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.223664} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.059019] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 950.059019] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Deleted contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 950.059019] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 950.059019] env[62965]: INFO nova.compute.manager [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Took 2.34 seconds to destroy the instance on the hypervisor. [ 950.059019] env[62965]: DEBUG oslo.service.loopingcall [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 950.059019] env[62965]: INFO nova.compute.manager [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Took 16.82 seconds to build instance. [ 950.060367] env[62965]: DEBUG nova.compute.manager [-] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 950.060367] env[62965]: DEBUG nova.network.neutron [-] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 950.086497] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e73a534f-af98-4393-85f0-5e6378f4083e tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Releasing lock "refresh_cache-d038608c-9264-4fd5-b9ba-0653c98ec0e4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.086653] env[62965]: DEBUG oslo_concurrency.lockutils [req-733fc234-16b9-4330-b6b5-186b70729774 req-a9f22bb2-2f8d-4482-9c5b-cfbba395ea8e service nova] Acquired lock "refresh_cache-d038608c-9264-4fd5-b9ba-0653c98ec0e4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.086849] env[62965]: DEBUG nova.network.neutron [req-733fc234-16b9-4330-b6b5-186b70729774 req-a9f22bb2-2f8d-4482-9c5b-cfbba395ea8e service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Refreshing network info cache for port b1718061-edf1-4077-b57f-0e54aea365a6 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 950.093246] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Releasing lock "refresh_cache-c2448b76-1553-4b68-a731-0bb0967f4c1d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.093246] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Updated the network info_cache for instance {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10328}} [ 950.093246] env[62965]: DEBUG oslo_concurrency.lockutils [None req-085cd9fc-7ae6-490b-8cdf-8365bf1ee3c0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquired lock "refresh_cache-c2448b76-1553-4b68-a731-0bb0967f4c1d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.093246] env[62965]: DEBUG nova.network.neutron [None req-085cd9fc-7ae6-490b-8cdf-8365bf1ee3c0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 950.094145] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 950.094412] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 950.094615] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 950.094809] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 950.095008] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 950.095203] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 950.095555] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62965) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10876}} [ 950.095555] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 950.187746] env[62965]: DEBUG oslo_vmware.api [None req-f99d4b6d-896b-445f-bf43-4d0b2692cbd7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390119, 'name': CloneVM_Task} progress is 94%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.214965] env[62965]: DEBUG nova.scheduler.client.report [None req-a4245b41-dd2b-44aa-b46d-f458a24a3a34 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 950.445034] env[62965]: DEBUG nova.compute.manager [req-ee947a2c-4cdc-4c68-9e68-0493a035dbee req-8c11eaa9-0bc7-4c4d-a3e1-78c7ce2e282e service nova] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Received event network-vif-deleted-c6f37c93-3ad8-4408-9f81-519246ae7051 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 950.445291] env[62965]: INFO nova.compute.manager [req-ee947a2c-4cdc-4c68-9e68-0493a035dbee req-8c11eaa9-0bc7-4c4d-a3e1-78c7ce2e282e service nova] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Neutron deleted interface c6f37c93-3ad8-4408-9f81-519246ae7051; detaching it from the instance and deleting it from the info cache [ 950.445511] env[62965]: DEBUG nova.network.neutron [req-ee947a2c-4cdc-4c68-9e68-0493a035dbee req-8c11eaa9-0bc7-4c4d-a3e1-78c7ce2e282e service nova] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.446757] env[62965]: WARNING oslo_messaging._drivers.amqpdriver [req-ee947a2c-4cdc-4c68-9e68-0493a035dbee req-8c11eaa9-0bc7-4c4d-a3e1-78c7ce2e282e service nova] Number of call queues is 11, greater than warning threshold: 10. There could be a leak. Increasing threshold to: 20 [ 950.563321] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9b60f0f6-bee1-4e2a-820f-1cfcf088de50 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Lock "3223b1ed-abc1-439c-bcad-6317eb204e45" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.335s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.572111] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b54f7baf-59b0-4c6b-9912-8ff68c41ea55 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "interface-00f24976-2f0e-4a2f-8f2e-012b0fce479f-9c46a8bf-6b23-45d3-95a4-f0c751895c71" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.572407] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b54f7baf-59b0-4c6b-9912-8ff68c41ea55 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "interface-00f24976-2f0e-4a2f-8f2e-012b0fce479f-9c46a8bf-6b23-45d3-95a4-f0c751895c71" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.572931] env[62965]: DEBUG nova.objects.instance [None req-b54f7baf-59b0-4c6b-9912-8ff68c41ea55 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lazy-loading 'flavor' on Instance uuid 00f24976-2f0e-4a2f-8f2e-012b0fce479f {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 950.591894] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e73a534f-af98-4393-85f0-5e6378f4083e tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "interface-d038608c-9264-4fd5-b9ba-0653c98ec0e4-9c46a8bf-6b23-45d3-95a4-f0c751895c71" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.087s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.599068] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.688173] env[62965]: DEBUG oslo_vmware.api [None req-f99d4b6d-896b-445f-bf43-4d0b2692cbd7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390119, 'name': CloneVM_Task} progress is 95%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.692830] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9ec520c8-ed90-4e25-a162-773f48db0ed6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Acquiring lock "3223b1ed-abc1-439c-bcad-6317eb204e45" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.693207] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9ec520c8-ed90-4e25-a162-773f48db0ed6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Lock "3223b1ed-abc1-439c-bcad-6317eb204e45" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.693399] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9ec520c8-ed90-4e25-a162-773f48db0ed6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Acquiring lock "3223b1ed-abc1-439c-bcad-6317eb204e45-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.693648] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9ec520c8-ed90-4e25-a162-773f48db0ed6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Lock "3223b1ed-abc1-439c-bcad-6317eb204e45-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.693931] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9ec520c8-ed90-4e25-a162-773f48db0ed6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Lock "3223b1ed-abc1-439c-bcad-6317eb204e45-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.696584] env[62965]: INFO nova.compute.manager [None req-9ec520c8-ed90-4e25-a162-773f48db0ed6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Terminating instance [ 950.859088] env[62965]: DEBUG nova.compute.manager [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 950.891013] env[62965]: DEBUG nova.virt.hardware [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 950.891013] env[62965]: DEBUG nova.virt.hardware [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 950.891013] env[62965]: DEBUG nova.virt.hardware [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 950.891013] env[62965]: DEBUG nova.virt.hardware [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 950.891013] env[62965]: DEBUG nova.virt.hardware [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 950.891347] env[62965]: DEBUG nova.virt.hardware [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 950.891400] env[62965]: DEBUG nova.virt.hardware [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 950.891521] env[62965]: DEBUG nova.virt.hardware [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 950.891757] env[62965]: DEBUG nova.virt.hardware [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 950.891853] env[62965]: DEBUG nova.virt.hardware [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 950.892038] env[62965]: DEBUG nova.virt.hardware [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 950.892921] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53b74b50-c5df-42b3-a4f2-784c814929a2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.902770] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-471debd6-72b7-45b0-ad6f-cc819486ee70 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.907526] env[62965]: DEBUG nova.network.neutron [-] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.941329] env[62965]: DEBUG nova.network.neutron [req-733fc234-16b9-4330-b6b5-186b70729774 req-a9f22bb2-2f8d-4482-9c5b-cfbba395ea8e service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Updated VIF entry in instance network info cache for port b1718061-edf1-4077-b57f-0e54aea365a6. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 950.941716] env[62965]: DEBUG nova.network.neutron [req-733fc234-16b9-4330-b6b5-186b70729774 req-a9f22bb2-2f8d-4482-9c5b-cfbba395ea8e service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Updating instance_info_cache with network_info: [{"id": "b1718061-edf1-4077-b57f-0e54aea365a6", "address": "fa:16:3e:1f:27:16", "network": {"id": "a9699d8b-5e76-4666-a236-33ab8f22345e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1329540850-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1952dedf340f4b7ba0d57af6bba9a749", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a316376e-2ef0-4b1e-b40c-10321ebd7e1a", "external-id": "nsx-vlan-transportzone-942", "segmentation_id": 942, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb1718061-ed", "ovs_interfaceid": "b1718061-edf1-4077-b57f-0e54aea365a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.950448] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-06e93430-8fec-4013-9fc8-cbfaeefdd8c4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.960404] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9704404d-d0e3-41ab-8384-af36b3f09f1e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.972518] env[62965]: DEBUG nova.network.neutron [None req-085cd9fc-7ae6-490b-8cdf-8365bf1ee3c0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Updating instance_info_cache with network_info: [{"id": "57672e8b-1df1-463d-b9b3-8290667d9fd2", "address": "fa:16:3e:53:46:47", "network": {"id": "4649fed7-c2ca-43a9-a183-7c99423da726", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1978698173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fab418e2fe1420793517663574b43bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43ad01d2-c7dd-453c-a929-8ad76294d13c", "external-id": "nsx-vlan-transportzone-176", "segmentation_id": 176, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57672e8b-1d", "ovs_interfaceid": "57672e8b-1df1-463d-b9b3-8290667d9fd2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.999896] env[62965]: DEBUG nova.compute.manager [req-ee947a2c-4cdc-4c68-9e68-0493a035dbee req-8c11eaa9-0bc7-4c4d-a3e1-78c7ce2e282e service nova] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Detach interface failed, port_id=c6f37c93-3ad8-4408-9f81-519246ae7051, reason: Instance 0eb943ae-462c-40dd-a870-7a884ea2db74 could not be found. {{(pid=62965) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11294}} [ 951.174766] env[62965]: DEBUG nova.objects.instance [None req-b54f7baf-59b0-4c6b-9912-8ff68c41ea55 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lazy-loading 'pci_requests' on Instance uuid 00f24976-2f0e-4a2f-8f2e-012b0fce479f {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 951.188206] env[62965]: DEBUG oslo_vmware.api [None req-f99d4b6d-896b-445f-bf43-4d0b2692cbd7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390119, 'name': CloneVM_Task, 'duration_secs': 1.881643} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.188497] env[62965]: INFO nova.virt.vmwareapi.vmops [None req-f99d4b6d-896b-445f-bf43-4d0b2692cbd7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Created linked-clone VM from snapshot [ 951.189251] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f48497fb-1898-4f81-a1f3-f65c7e9bb7ad {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.200479] env[62965]: DEBUG nova.compute.manager [None req-9ec520c8-ed90-4e25-a162-773f48db0ed6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 951.200702] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-9ec520c8-ed90-4e25-a162-773f48db0ed6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 951.200962] env[62965]: DEBUG nova.virt.vmwareapi.images [None req-f99d4b6d-896b-445f-bf43-4d0b2692cbd7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Uploading image d9a5bf3d-acd8-42c1-a30e-57bff64aa7ed {{(pid=62965) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 951.203525] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-925c6c3f-1e15-4224-9ea2-bf795be7b6e1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.211067] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ec520c8-ed90-4e25-a162-773f48db0ed6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 951.211329] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a34f7fd3-9ae3-45f1-901c-487c945bdab1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.213615] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-f99d4b6d-896b-445f-bf43-4d0b2692cbd7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Destroying the VM {{(pid=62965) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 951.214102] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-75cff535-9bee-492f-a9a3-a163431a7465 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.220042] env[62965]: DEBUG oslo_vmware.api [None req-9ec520c8-ed90-4e25-a162-773f48db0ed6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 951.220042] env[62965]: value = "task-1390123" [ 951.220042] env[62965]: _type = "Task" [ 951.220042] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.225363] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a4245b41-dd2b-44aa-b46d-f458a24a3a34 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.394s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.230300] env[62965]: DEBUG oslo_vmware.api [None req-f99d4b6d-896b-445f-bf43-4d0b2692cbd7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 951.230300] env[62965]: value = "task-1390124" [ 951.230300] env[62965]: _type = "Task" [ 951.230300] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.231304] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.346s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.233261] env[62965]: INFO nova.compute.claims [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 951.243959] env[62965]: DEBUG oslo_vmware.api [None req-9ec520c8-ed90-4e25-a162-773f48db0ed6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390123, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.250386] env[62965]: DEBUG oslo_vmware.api [None req-f99d4b6d-896b-445f-bf43-4d0b2692cbd7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390124, 'name': Destroy_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.410828] env[62965]: INFO nova.compute.manager [-] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Took 1.35 seconds to deallocate network for instance. [ 951.445399] env[62965]: DEBUG oslo_concurrency.lockutils [req-733fc234-16b9-4330-b6b5-186b70729774 req-a9f22bb2-2f8d-4482-9c5b-cfbba395ea8e service nova] Releasing lock "refresh_cache-d038608c-9264-4fd5-b9ba-0653c98ec0e4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.479036] env[62965]: DEBUG oslo_concurrency.lockutils [None req-085cd9fc-7ae6-490b-8cdf-8365bf1ee3c0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Releasing lock "refresh_cache-c2448b76-1553-4b68-a731-0bb0967f4c1d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.481898] env[62965]: DEBUG nova.compute.manager [None req-085cd9fc-7ae6-490b-8cdf-8365bf1ee3c0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 951.483045] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8f784e7-7371-4f34-b9c8-db95d2b24392 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.628229] env[62965]: DEBUG nova.network.neutron [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Successfully updated port: 219e79ce-5ab0-49eb-a38f-6b1ba814cdf3 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 951.678272] env[62965]: DEBUG nova.objects.base [None req-b54f7baf-59b0-4c6b-9912-8ff68c41ea55 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Object Instance<00f24976-2f0e-4a2f-8f2e-012b0fce479f> lazy-loaded attributes: flavor,pci_requests {{(pid=62965) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 951.678502] env[62965]: DEBUG nova.network.neutron [None req-b54f7baf-59b0-4c6b-9912-8ff68c41ea55 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 951.743498] env[62965]: DEBUG oslo_vmware.api [None req-9ec520c8-ed90-4e25-a162-773f48db0ed6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390123, 'name': PowerOffVM_Task, 'duration_secs': 0.246717} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.750318] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ec520c8-ed90-4e25-a162-773f48db0ed6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 951.750318] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-9ec520c8-ed90-4e25-a162-773f48db0ed6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 951.751782] env[62965]: DEBUG nova.policy [None req-b54f7baf-59b0-4c6b-9912-8ff68c41ea55 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '182d38ca18c64bed8f3ecc3f95229756', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1952dedf340f4b7ba0d57af6bba9a749', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 951.760825] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-81af448c-60f6-48fa-a86c-70401a597952 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.778154] env[62965]: DEBUG oslo_vmware.api [None req-f99d4b6d-896b-445f-bf43-4d0b2692cbd7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390124, 'name': Destroy_Task} progress is 33%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.817895] env[62965]: INFO nova.scheduler.client.report [None req-a4245b41-dd2b-44aa-b46d-f458a24a3a34 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Deleted allocation for migration 5db4bd92-be48-40f0-88ac-7cda961107d8 [ 951.846178] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-9ec520c8-ed90-4e25-a162-773f48db0ed6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 951.846626] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-9ec520c8-ed90-4e25-a162-773f48db0ed6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Deleting contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 951.846961] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ec520c8-ed90-4e25-a162-773f48db0ed6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Deleting the datastore file [datastore2] 3223b1ed-abc1-439c-bcad-6317eb204e45 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 951.847428] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-90298dfc-d13b-41a4-98bf-0c49ef1e9f36 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.858450] env[62965]: DEBUG oslo_vmware.api [None req-9ec520c8-ed90-4e25-a162-773f48db0ed6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for the task: (returnval){ [ 951.858450] env[62965]: value = "task-1390126" [ 951.858450] env[62965]: _type = "Task" [ 951.858450] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.870317] env[62965]: DEBUG oslo_vmware.api [None req-9ec520c8-ed90-4e25-a162-773f48db0ed6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390126, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.984244] env[62965]: INFO nova.compute.manager [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Took 0.57 seconds to detach 1 volumes for instance. [ 952.132121] env[62965]: DEBUG oslo_concurrency.lockutils [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquiring lock "refresh_cache-a63b301d-8d52-4fec-951f-a15cda9449ac" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.132216] env[62965]: DEBUG oslo_concurrency.lockutils [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquired lock "refresh_cache-a63b301d-8d52-4fec-951f-a15cda9449ac" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.132436] env[62965]: DEBUG nova.network.neutron [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 952.207563] env[62965]: DEBUG oslo_vmware.rw_handles [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5205ac28-03ec-e1dd-1f2d-02c5b4ff905d/disk-0.vmdk. {{(pid=62965) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 952.209197] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ed4eca2-042c-4634-b579-f12dd2f7d9ed {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.216264] env[62965]: DEBUG nova.compute.manager [req-499ab959-eb1a-4b0e-8529-79935d7f530b req-1506421c-b5ba-4a65-8c82-14989ac3bcfd service nova] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Received event network-changed-8fea9d40-d98f-4328-ad3e-ae708a849908 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 952.216560] env[62965]: DEBUG nova.compute.manager [req-499ab959-eb1a-4b0e-8529-79935d7f530b req-1506421c-b5ba-4a65-8c82-14989ac3bcfd service nova] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Refreshing instance network info cache due to event network-changed-8fea9d40-d98f-4328-ad3e-ae708a849908. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 952.216861] env[62965]: DEBUG oslo_concurrency.lockutils [req-499ab959-eb1a-4b0e-8529-79935d7f530b req-1506421c-b5ba-4a65-8c82-14989ac3bcfd service nova] Acquiring lock "refresh_cache-00f24976-2f0e-4a2f-8f2e-012b0fce479f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.217078] env[62965]: DEBUG oslo_concurrency.lockutils [req-499ab959-eb1a-4b0e-8529-79935d7f530b req-1506421c-b5ba-4a65-8c82-14989ac3bcfd service nova] Acquired lock "refresh_cache-00f24976-2f0e-4a2f-8f2e-012b0fce479f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.217530] env[62965]: DEBUG nova.network.neutron [req-499ab959-eb1a-4b0e-8529-79935d7f530b req-1506421c-b5ba-4a65-8c82-14989ac3bcfd service nova] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Refreshing network info cache for port 8fea9d40-d98f-4328-ad3e-ae708a849908 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 952.221641] env[62965]: DEBUG oslo_vmware.rw_handles [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5205ac28-03ec-e1dd-1f2d-02c5b4ff905d/disk-0.vmdk is in state: ready. {{(pid=62965) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 952.221964] env[62965]: ERROR oslo_vmware.rw_handles [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5205ac28-03ec-e1dd-1f2d-02c5b4ff905d/disk-0.vmdk due to incomplete transfer. [ 952.222424] env[62965]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-7a805196-62c3-499c-a0e2-9ef9800b82c5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.232801] env[62965]: DEBUG oslo_vmware.rw_handles [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5205ac28-03ec-e1dd-1f2d-02c5b4ff905d/disk-0.vmdk. {{(pid=62965) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 952.233521] env[62965]: DEBUG nova.virt.vmwareapi.images [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Uploaded image c1230705-5bdc-4c15-a740-13db3313def2 to the Glance image server {{(pid=62965) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 952.236983] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Destroying the VM {{(pid=62965) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 952.237866] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-b80cc89d-f756-46a9-a04d-5e0e5c095243 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.246948] env[62965]: DEBUG oslo_vmware.api [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 952.246948] env[62965]: value = "task-1390127" [ 952.246948] env[62965]: _type = "Task" [ 952.246948] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.269625] env[62965]: DEBUG oslo_vmware.api [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390127, 'name': Destroy_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.275158] env[62965]: DEBUG oslo_vmware.api [None req-f99d4b6d-896b-445f-bf43-4d0b2692cbd7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390124, 'name': Destroy_Task, 'duration_secs': 0.739164} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.276213] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-f99d4b6d-896b-445f-bf43-4d0b2692cbd7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Destroyed the VM [ 952.276827] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-f99d4b6d-896b-445f-bf43-4d0b2692cbd7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Deleting Snapshot of the VM instance {{(pid=62965) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 952.280171] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-abf00358-39a4-425b-9a27-d29fcbdd44c4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.294566] env[62965]: DEBUG oslo_vmware.api [None req-f99d4b6d-896b-445f-bf43-4d0b2692cbd7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 952.294566] env[62965]: value = "task-1390128" [ 952.294566] env[62965]: _type = "Task" [ 952.294566] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.311905] env[62965]: DEBUG oslo_vmware.api [None req-f99d4b6d-896b-445f-bf43-4d0b2692cbd7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390128, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.323950] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a4245b41-dd2b-44aa-b46d-f458a24a3a34 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "fbbcbe60-92f3-497d-b190-520a2547b3a9" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 9.238s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.373520] env[62965]: DEBUG oslo_vmware.api [None req-9ec520c8-ed90-4e25-a162-773f48db0ed6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Task: {'id': task-1390126, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.314599} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.373778] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ec520c8-ed90-4e25-a162-773f48db0ed6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 952.374418] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-9ec520c8-ed90-4e25-a162-773f48db0ed6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Deleted contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 952.374418] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-9ec520c8-ed90-4e25-a162-773f48db0ed6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 952.374565] env[62965]: INFO nova.compute.manager [None req-9ec520c8-ed90-4e25-a162-773f48db0ed6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Took 1.17 seconds to destroy the instance on the hypervisor. [ 952.374874] env[62965]: DEBUG oslo.service.loopingcall [None req-9ec520c8-ed90-4e25-a162-773f48db0ed6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 952.375149] env[62965]: DEBUG nova.compute.manager [-] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 952.375374] env[62965]: DEBUG nova.network.neutron [-] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 952.491653] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.496527] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d361627-2a20-4676-8079-4ef6e0d854d3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.505661] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88c22999-63e1-4a76-98a7-bffb9e570a6e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.512800] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a952bce5-9c5f-4269-b506-2e22526136bc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.517741] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-085cd9fc-7ae6-490b-8cdf-8365bf1ee3c0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Doing hard reboot of VM {{(pid=62965) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 952.518483] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-a4cb7bff-9495-4bdb-bc5a-b6d0289026a1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.552012] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a1854a2-5428-4ef2-a839-5df8a1f0a77a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.554657] env[62965]: DEBUG oslo_vmware.api [None req-085cd9fc-7ae6-490b-8cdf-8365bf1ee3c0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 952.554657] env[62965]: value = "task-1390129" [ 952.554657] env[62965]: _type = "Task" [ 952.554657] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.563187] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec290be4-f1d7-4958-a2a1-9c5f9e3d4145 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.572223] env[62965]: DEBUG oslo_vmware.api [None req-085cd9fc-7ae6-490b-8cdf-8365bf1ee3c0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390129, 'name': ResetVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.585181] env[62965]: DEBUG nova.compute.provider_tree [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 952.593393] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.696952] env[62965]: DEBUG nova.network.neutron [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 952.760486] env[62965]: DEBUG oslo_vmware.api [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390127, 'name': Destroy_Task} progress is 33%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.779032] env[62965]: DEBUG nova.compute.manager [req-84a0c211-7809-497d-8947-76794466bfc6 req-347a5720-b450-4b17-90f2-58ba1f60cd68 service nova] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Received event network-vif-plugged-219e79ce-5ab0-49eb-a38f-6b1ba814cdf3 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 952.779220] env[62965]: DEBUG oslo_concurrency.lockutils [req-84a0c211-7809-497d-8947-76794466bfc6 req-347a5720-b450-4b17-90f2-58ba1f60cd68 service nova] Acquiring lock "a63b301d-8d52-4fec-951f-a15cda9449ac-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.779471] env[62965]: DEBUG oslo_concurrency.lockutils [req-84a0c211-7809-497d-8947-76794466bfc6 req-347a5720-b450-4b17-90f2-58ba1f60cd68 service nova] Lock "a63b301d-8d52-4fec-951f-a15cda9449ac-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.779645] env[62965]: DEBUG oslo_concurrency.lockutils [req-84a0c211-7809-497d-8947-76794466bfc6 req-347a5720-b450-4b17-90f2-58ba1f60cd68 service nova] Lock "a63b301d-8d52-4fec-951f-a15cda9449ac-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.779819] env[62965]: DEBUG nova.compute.manager [req-84a0c211-7809-497d-8947-76794466bfc6 req-347a5720-b450-4b17-90f2-58ba1f60cd68 service nova] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] No waiting events found dispatching network-vif-plugged-219e79ce-5ab0-49eb-a38f-6b1ba814cdf3 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 952.779991] env[62965]: WARNING nova.compute.manager [req-84a0c211-7809-497d-8947-76794466bfc6 req-347a5720-b450-4b17-90f2-58ba1f60cd68 service nova] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Received unexpected event network-vif-plugged-219e79ce-5ab0-49eb-a38f-6b1ba814cdf3 for instance with vm_state building and task_state spawning. [ 952.780348] env[62965]: DEBUG nova.compute.manager [req-84a0c211-7809-497d-8947-76794466bfc6 req-347a5720-b450-4b17-90f2-58ba1f60cd68 service nova] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Received event network-changed-219e79ce-5ab0-49eb-a38f-6b1ba814cdf3 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 952.780485] env[62965]: DEBUG nova.compute.manager [req-84a0c211-7809-497d-8947-76794466bfc6 req-347a5720-b450-4b17-90f2-58ba1f60cd68 service nova] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Refreshing instance network info cache due to event network-changed-219e79ce-5ab0-49eb-a38f-6b1ba814cdf3. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 952.780576] env[62965]: DEBUG oslo_concurrency.lockutils [req-84a0c211-7809-497d-8947-76794466bfc6 req-347a5720-b450-4b17-90f2-58ba1f60cd68 service nova] Acquiring lock "refresh_cache-a63b301d-8d52-4fec-951f-a15cda9449ac" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.807025] env[62965]: DEBUG oslo_vmware.api [None req-f99d4b6d-896b-445f-bf43-4d0b2692cbd7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390128, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.000398] env[62965]: DEBUG nova.network.neutron [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Updating instance_info_cache with network_info: [{"id": "219e79ce-5ab0-49eb-a38f-6b1ba814cdf3", "address": "fa:16:3e:6d:19:17", "network": {"id": "98b821d6-efa4-46c8-81e6-a703bd34cb6a", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1649479803-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a539017d4b84f39a71b6f4098395fbe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c979f78-8597-41f8-b1de-995014032689", "external-id": "nsx-vlan-transportzone-477", "segmentation_id": 477, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap219e79ce-5a", "ovs_interfaceid": "219e79ce-5ab0-49eb-a38f-6b1ba814cdf3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.066847] env[62965]: DEBUG oslo_vmware.api [None req-085cd9fc-7ae6-490b-8cdf-8365bf1ee3c0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390129, 'name': ResetVM_Task, 'duration_secs': 0.127985} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.069502] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-085cd9fc-7ae6-490b-8cdf-8365bf1ee3c0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Did hard reboot of VM {{(pid=62965) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 953.069703] env[62965]: DEBUG nova.compute.manager [None req-085cd9fc-7ae6-490b-8cdf-8365bf1ee3c0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 953.070930] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc83761f-cfed-4db5-bef0-b597b4fd3163 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.088859] env[62965]: DEBUG nova.scheduler.client.report [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 953.156543] env[62965]: DEBUG nova.network.neutron [req-499ab959-eb1a-4b0e-8529-79935d7f530b req-1506421c-b5ba-4a65-8c82-14989ac3bcfd service nova] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Updated VIF entry in instance network info cache for port 8fea9d40-d98f-4328-ad3e-ae708a849908. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 953.156543] env[62965]: DEBUG nova.network.neutron [req-499ab959-eb1a-4b0e-8529-79935d7f530b req-1506421c-b5ba-4a65-8c82-14989ac3bcfd service nova] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Updating instance_info_cache with network_info: [{"id": "8fea9d40-d98f-4328-ad3e-ae708a849908", "address": "fa:16:3e:c3:8c:35", "network": {"id": "a9699d8b-5e76-4666-a236-33ab8f22345e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1329540850-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1952dedf340f4b7ba0d57af6bba9a749", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a316376e-2ef0-4b1e-b40c-10321ebd7e1a", "external-id": "nsx-vlan-transportzone-942", "segmentation_id": 942, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8fea9d40-d9", "ovs_interfaceid": "8fea9d40-d98f-4328-ad3e-ae708a849908", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.258748] env[62965]: DEBUG oslo_vmware.api [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390127, 'name': Destroy_Task, 'duration_secs': 0.917773} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.259590] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Destroyed the VM [ 953.259937] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Deleting Snapshot of the VM instance {{(pid=62965) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 953.260302] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-33c6e2a5-495d-4ca3-b8dd-ed78a0ff16a9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.268317] env[62965]: DEBUG oslo_vmware.api [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 953.268317] env[62965]: value = "task-1390130" [ 953.268317] env[62965]: _type = "Task" [ 953.268317] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.278744] env[62965]: DEBUG oslo_vmware.api [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390130, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.307346] env[62965]: DEBUG oslo_vmware.api [None req-f99d4b6d-896b-445f-bf43-4d0b2692cbd7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390128, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.390772] env[62965]: DEBUG nova.network.neutron [None req-b54f7baf-59b0-4c6b-9912-8ff68c41ea55 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Successfully updated port: 9c46a8bf-6b23-45d3-95a4-f0c751895c71 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 953.430627] env[62965]: DEBUG nova.network.neutron [-] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.503981] env[62965]: DEBUG oslo_concurrency.lockutils [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Releasing lock "refresh_cache-a63b301d-8d52-4fec-951f-a15cda9449ac" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.504285] env[62965]: DEBUG nova.compute.manager [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Instance network_info: |[{"id": "219e79ce-5ab0-49eb-a38f-6b1ba814cdf3", "address": "fa:16:3e:6d:19:17", "network": {"id": "98b821d6-efa4-46c8-81e6-a703bd34cb6a", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1649479803-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a539017d4b84f39a71b6f4098395fbe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c979f78-8597-41f8-b1de-995014032689", "external-id": "nsx-vlan-transportzone-477", "segmentation_id": 477, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap219e79ce-5a", "ovs_interfaceid": "219e79ce-5ab0-49eb-a38f-6b1ba814cdf3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 953.504655] env[62965]: DEBUG oslo_concurrency.lockutils [req-84a0c211-7809-497d-8947-76794466bfc6 req-347a5720-b450-4b17-90f2-58ba1f60cd68 service nova] Acquired lock "refresh_cache-a63b301d-8d52-4fec-951f-a15cda9449ac" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.504853] env[62965]: DEBUG nova.network.neutron [req-84a0c211-7809-497d-8947-76794466bfc6 req-347a5720-b450-4b17-90f2-58ba1f60cd68 service nova] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Refreshing network info cache for port 219e79ce-5ab0-49eb-a38f-6b1ba814cdf3 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 953.506192] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6d:19:17', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8c979f78-8597-41f8-b1de-995014032689', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '219e79ce-5ab0-49eb-a38f-6b1ba814cdf3', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 953.515488] env[62965]: DEBUG oslo.service.loopingcall [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 953.516831] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 953.517120] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-684af1ed-bf75-44f9-a679-ee188d10880f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.541112] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 953.541112] env[62965]: value = "task-1390131" [ 953.541112] env[62965]: _type = "Task" [ 953.541112] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.557614] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390131, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.590443] env[62965]: DEBUG oslo_concurrency.lockutils [None req-085cd9fc-7ae6-490b-8cdf-8365bf1ee3c0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "c2448b76-1553-4b68-a731-0bb0967f4c1d" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 6.110s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.594846] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.364s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.595376] env[62965]: DEBUG nova.compute.manager [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 953.597894] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 2.999s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.598474] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.598701] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62965) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 953.599100] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.108s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.599590] env[62965]: DEBUG nova.objects.instance [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lazy-loading 'resources' on Instance uuid 0eb943ae-462c-40dd-a870-7a884ea2db74 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 953.602146] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74c275d6-8cd4-4e67-9396-1634b24bc34c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.616947] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-854f2aca-2bcf-465a-99a4-c95c92a9bf4e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.637997] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ce05d90-784c-4bdf-8138-f1197f1aef1b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.647923] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5bd7a35-7bdd-412a-8c4c-dbb21f13ceaf {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.682665] env[62965]: DEBUG oslo_concurrency.lockutils [req-499ab959-eb1a-4b0e-8529-79935d7f530b req-1506421c-b5ba-4a65-8c82-14989ac3bcfd service nova] Releasing lock "refresh_cache-00f24976-2f0e-4a2f-8f2e-012b0fce479f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.683595] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179883MB free_disk=174GB free_vcpus=48 pci_devices=None {{(pid=62965) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 953.683738] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.780234] env[62965]: DEBUG oslo_vmware.api [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390130, 'name': RemoveSnapshot_Task, 'duration_secs': 0.395688} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.780517] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Deleted Snapshot of the VM instance {{(pid=62965) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 953.780797] env[62965]: DEBUG nova.compute.manager [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 953.781625] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1b60863-62eb-4c38-bc2a-fd5a13dd5717 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.807234] env[62965]: DEBUG oslo_vmware.api [None req-f99d4b6d-896b-445f-bf43-4d0b2692cbd7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390128, 'name': RemoveSnapshot_Task, 'duration_secs': 1.227083} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.811046] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-f99d4b6d-896b-445f-bf43-4d0b2692cbd7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Deleted Snapshot of the VM instance {{(pid=62965) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 953.893398] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b54f7baf-59b0-4c6b-9912-8ff68c41ea55 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "refresh_cache-00f24976-2f0e-4a2f-8f2e-012b0fce479f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.893646] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b54f7baf-59b0-4c6b-9912-8ff68c41ea55 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquired lock "refresh_cache-00f24976-2f0e-4a2f-8f2e-012b0fce479f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.893888] env[62965]: DEBUG nova.network.neutron [None req-b54f7baf-59b0-4c6b-9912-8ff68c41ea55 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 953.936915] env[62965]: INFO nova.compute.manager [-] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Took 1.56 seconds to deallocate network for instance. [ 954.053015] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390131, 'name': CreateVM_Task} progress is 25%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.102969] env[62965]: DEBUG nova.compute.utils [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 954.108023] env[62965]: DEBUG nova.compute.manager [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 954.108023] env[62965]: DEBUG nova.network.neutron [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 954.153175] env[62965]: DEBUG nova.policy [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '97ee4fdb7cdb45829c7fe694e883da8c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1ed71548c085499981fa1b102d660368', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 954.299469] env[62965]: INFO nova.compute.manager [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Shelve offloading [ 954.312554] env[62965]: WARNING nova.compute.manager [None req-f99d4b6d-896b-445f-bf43-4d0b2692cbd7 tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Image not found during snapshot: nova.exception.ImageNotFound: Image d9a5bf3d-acd8-42c1-a30e-57bff64aa7ed could not be found. [ 954.354423] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4332d9b6-cb0d-400a-ae00-a42d45deec4d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.366232] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-274beab6-e6bc-475a-b700-c73abce171a2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.405297] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ced2eaab-ca75-469b-8e52-3544c9a6fa32 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.414173] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1df4f0a9-380f-4f7a-8976-74ca316d2c04 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.432096] env[62965]: DEBUG nova.compute.provider_tree [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 954.443432] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9ec520c8-ed90-4e25-a162-773f48db0ed6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.464198] env[62965]: WARNING nova.network.neutron [None req-b54f7baf-59b0-4c6b-9912-8ff68c41ea55 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] a9699d8b-5e76-4666-a236-33ab8f22345e already exists in list: networks containing: ['a9699d8b-5e76-4666-a236-33ab8f22345e']. ignoring it [ 954.466884] env[62965]: DEBUG nova.network.neutron [req-84a0c211-7809-497d-8947-76794466bfc6 req-347a5720-b450-4b17-90f2-58ba1f60cd68 service nova] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Updated VIF entry in instance network info cache for port 219e79ce-5ab0-49eb-a38f-6b1ba814cdf3. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 954.467588] env[62965]: DEBUG nova.network.neutron [req-84a0c211-7809-497d-8947-76794466bfc6 req-347a5720-b450-4b17-90f2-58ba1f60cd68 service nova] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Updating instance_info_cache with network_info: [{"id": "219e79ce-5ab0-49eb-a38f-6b1ba814cdf3", "address": "fa:16:3e:6d:19:17", "network": {"id": "98b821d6-efa4-46c8-81e6-a703bd34cb6a", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1649479803-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a539017d4b84f39a71b6f4098395fbe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c979f78-8597-41f8-b1de-995014032689", "external-id": "nsx-vlan-transportzone-477", "segmentation_id": 477, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap219e79ce-5a", "ovs_interfaceid": "219e79ce-5ab0-49eb-a38f-6b1ba814cdf3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.512302] env[62965]: DEBUG nova.network.neutron [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Successfully created port: 6b5a028e-baa2-4d3d-a9ca-ade1c685e4b9 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 954.554402] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390131, 'name': CreateVM_Task, 'duration_secs': 0.76285} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.554586] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 954.555382] env[62965]: DEBUG oslo_concurrency.lockutils [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.555497] env[62965]: DEBUG oslo_concurrency.lockutils [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.555776] env[62965]: DEBUG oslo_concurrency.lockutils [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 954.556053] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-42c986be-c8ab-41dc-bf7d-93acd5a98fd7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.561961] env[62965]: DEBUG oslo_vmware.api [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Waiting for the task: (returnval){ [ 954.561961] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52993ffd-5b14-00f0-80b1-e904baac699b" [ 954.561961] env[62965]: _type = "Task" [ 954.561961] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.571501] env[62965]: DEBUG oslo_vmware.api [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52993ffd-5b14-00f0-80b1-e904baac699b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.607957] env[62965]: DEBUG nova.compute.manager [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 954.619779] env[62965]: DEBUG oslo_concurrency.lockutils [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquiring lock "97954de1-7b4b-4615-8d0f-0e8580a515ed" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.619779] env[62965]: DEBUG oslo_concurrency.lockutils [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "97954de1-7b4b-4615-8d0f-0e8580a515ed" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.619779] env[62965]: DEBUG oslo_concurrency.lockutils [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquiring lock "97954de1-7b4b-4615-8d0f-0e8580a515ed-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.619779] env[62965]: DEBUG oslo_concurrency.lockutils [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "97954de1-7b4b-4615-8d0f-0e8580a515ed-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.619779] env[62965]: DEBUG oslo_concurrency.lockutils [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "97954de1-7b4b-4615-8d0f-0e8580a515ed-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.621746] env[62965]: INFO nova.compute.manager [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Terminating instance [ 954.808430] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 954.808871] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-77342c33-d14a-48c8-bb3e-763bd603144c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.818040] env[62965]: DEBUG oslo_vmware.api [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 954.818040] env[62965]: value = "task-1390132" [ 954.818040] env[62965]: _type = "Task" [ 954.818040] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.819095] env[62965]: DEBUG nova.network.neutron [None req-b54f7baf-59b0-4c6b-9912-8ff68c41ea55 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Updating instance_info_cache with network_info: [{"id": "8fea9d40-d98f-4328-ad3e-ae708a849908", "address": "fa:16:3e:c3:8c:35", "network": {"id": "a9699d8b-5e76-4666-a236-33ab8f22345e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1329540850-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1952dedf340f4b7ba0d57af6bba9a749", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a316376e-2ef0-4b1e-b40c-10321ebd7e1a", "external-id": "nsx-vlan-transportzone-942", "segmentation_id": 942, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8fea9d40-d9", "ovs_interfaceid": "8fea9d40-d98f-4328-ad3e-ae708a849908", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "9c46a8bf-6b23-45d3-95a4-f0c751895c71", "address": "fa:16:3e:36:f8:d3", "network": {"id": "a9699d8b-5e76-4666-a236-33ab8f22345e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1329540850-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1952dedf340f4b7ba0d57af6bba9a749", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a316376e-2ef0-4b1e-b40c-10321ebd7e1a", "external-id": "nsx-vlan-transportzone-942", "segmentation_id": 942, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c46a8bf-6b", "ovs_interfaceid": "9c46a8bf-6b23-45d3-95a4-f0c751895c71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.829948] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] VM already powered off {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 954.830301] env[62965]: DEBUG nova.compute.manager [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 954.831596] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57d6dbf0-49d2-434a-9af1-57a8f677711a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.838569] env[62965]: DEBUG oslo_concurrency.lockutils [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquiring lock "refresh_cache-a0e9cc1a-bc95-4a44-99c5-aabf85a373f4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.838740] env[62965]: DEBUG oslo_concurrency.lockutils [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquired lock "refresh_cache-a0e9cc1a-bc95-4a44-99c5-aabf85a373f4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.838912] env[62965]: DEBUG nova.network.neutron [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 954.935154] env[62965]: DEBUG nova.scheduler.client.report [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 954.970901] env[62965]: DEBUG oslo_concurrency.lockutils [req-84a0c211-7809-497d-8947-76794466bfc6 req-347a5720-b450-4b17-90f2-58ba1f60cd68 service nova] Releasing lock "refresh_cache-a63b301d-8d52-4fec-951f-a15cda9449ac" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.994356] env[62965]: DEBUG nova.compute.manager [req-1ecdf3f9-afc6-4d9f-b339-a014796be0e1 req-99ed7cee-0cc6-425d-96e5-c57bab759e71 service nova] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Received event network-vif-deleted-3eafc53f-b78d-43b3-9341-89b3a9cd0ecb {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 954.994549] env[62965]: DEBUG nova.compute.manager [req-1ecdf3f9-afc6-4d9f-b339-a014796be0e1 req-99ed7cee-0cc6-425d-96e5-c57bab759e71 service nova] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Received event network-vif-plugged-9c46a8bf-6b23-45d3-95a4-f0c751895c71 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 954.994701] env[62965]: DEBUG oslo_concurrency.lockutils [req-1ecdf3f9-afc6-4d9f-b339-a014796be0e1 req-99ed7cee-0cc6-425d-96e5-c57bab759e71 service nova] Acquiring lock "00f24976-2f0e-4a2f-8f2e-012b0fce479f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.996066] env[62965]: DEBUG oslo_concurrency.lockutils [req-1ecdf3f9-afc6-4d9f-b339-a014796be0e1 req-99ed7cee-0cc6-425d-96e5-c57bab759e71 service nova] Lock "00f24976-2f0e-4a2f-8f2e-012b0fce479f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.996294] env[62965]: DEBUG oslo_concurrency.lockutils [req-1ecdf3f9-afc6-4d9f-b339-a014796be0e1 req-99ed7cee-0cc6-425d-96e5-c57bab759e71 service nova] Lock "00f24976-2f0e-4a2f-8f2e-012b0fce479f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.996475] env[62965]: DEBUG nova.compute.manager [req-1ecdf3f9-afc6-4d9f-b339-a014796be0e1 req-99ed7cee-0cc6-425d-96e5-c57bab759e71 service nova] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] No waiting events found dispatching network-vif-plugged-9c46a8bf-6b23-45d3-95a4-f0c751895c71 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 954.996651] env[62965]: WARNING nova.compute.manager [req-1ecdf3f9-afc6-4d9f-b339-a014796be0e1 req-99ed7cee-0cc6-425d-96e5-c57bab759e71 service nova] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Received unexpected event network-vif-plugged-9c46a8bf-6b23-45d3-95a4-f0c751895c71 for instance with vm_state active and task_state None. [ 954.996819] env[62965]: DEBUG nova.compute.manager [req-1ecdf3f9-afc6-4d9f-b339-a014796be0e1 req-99ed7cee-0cc6-425d-96e5-c57bab759e71 service nova] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Received event network-changed-9c46a8bf-6b23-45d3-95a4-f0c751895c71 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 954.996983] env[62965]: DEBUG nova.compute.manager [req-1ecdf3f9-afc6-4d9f-b339-a014796be0e1 req-99ed7cee-0cc6-425d-96e5-c57bab759e71 service nova] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Refreshing instance network info cache due to event network-changed-9c46a8bf-6b23-45d3-95a4-f0c751895c71. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 954.997173] env[62965]: DEBUG oslo_concurrency.lockutils [req-1ecdf3f9-afc6-4d9f-b339-a014796be0e1 req-99ed7cee-0cc6-425d-96e5-c57bab759e71 service nova] Acquiring lock "refresh_cache-00f24976-2f0e-4a2f-8f2e-012b0fce479f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.072775] env[62965]: DEBUG oslo_vmware.api [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52993ffd-5b14-00f0-80b1-e904baac699b, 'name': SearchDatastore_Task, 'duration_secs': 0.012952} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.072775] env[62965]: DEBUG oslo_concurrency.lockutils [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.073020] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 955.073322] env[62965]: DEBUG oslo_concurrency.lockutils [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.073680] env[62965]: DEBUG oslo_concurrency.lockutils [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.073680] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 955.073847] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e6597d9b-87ff-4fb7-9bfe-56fb61e6e9c0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.083843] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 955.084097] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 955.084814] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f2cf3f5-96f1-4335-b130-c95ea46b46ab {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.090786] env[62965]: DEBUG oslo_vmware.api [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Waiting for the task: (returnval){ [ 955.090786] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52cccfd1-8eef-95de-cbce-413263a0eb0a" [ 955.090786] env[62965]: _type = "Task" [ 955.090786] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.098859] env[62965]: DEBUG oslo_vmware.api [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52cccfd1-8eef-95de-cbce-413263a0eb0a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.126922] env[62965]: DEBUG nova.compute.manager [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 955.127156] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 955.128349] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd9960aa-e2db-402a-ac79-c62eb759205e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.137614] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 955.137904] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f8d0d36c-9fb2-4a49-895b-833266c158b0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.144838] env[62965]: DEBUG oslo_vmware.api [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 955.144838] env[62965]: value = "task-1390133" [ 955.144838] env[62965]: _type = "Task" [ 955.144838] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.156817] env[62965]: DEBUG oslo_vmware.api [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390133, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.325047] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b54f7baf-59b0-4c6b-9912-8ff68c41ea55 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Releasing lock "refresh_cache-00f24976-2f0e-4a2f-8f2e-012b0fce479f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.325713] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b54f7baf-59b0-4c6b-9912-8ff68c41ea55 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "00f24976-2f0e-4a2f-8f2e-012b0fce479f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.325874] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b54f7baf-59b0-4c6b-9912-8ff68c41ea55 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquired lock "00f24976-2f0e-4a2f-8f2e-012b0fce479f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.326188] env[62965]: DEBUG oslo_concurrency.lockutils [req-1ecdf3f9-afc6-4d9f-b339-a014796be0e1 req-99ed7cee-0cc6-425d-96e5-c57bab759e71 service nova] Acquired lock "refresh_cache-00f24976-2f0e-4a2f-8f2e-012b0fce479f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.326394] env[62965]: DEBUG nova.network.neutron [req-1ecdf3f9-afc6-4d9f-b339-a014796be0e1 req-99ed7cee-0cc6-425d-96e5-c57bab759e71 service nova] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Refreshing network info cache for port 9c46a8bf-6b23-45d3-95a4-f0c751895c71 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 955.328448] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea5f4c30-eabb-4e88-89f0-b53c5ec6f5de {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.350269] env[62965]: DEBUG nova.virt.hardware [None req-b54f7baf-59b0-4c6b-9912-8ff68c41ea55 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 955.350526] env[62965]: DEBUG nova.virt.hardware [None req-b54f7baf-59b0-4c6b-9912-8ff68c41ea55 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 955.350779] env[62965]: DEBUG nova.virt.hardware [None req-b54f7baf-59b0-4c6b-9912-8ff68c41ea55 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 955.350967] env[62965]: DEBUG nova.virt.hardware [None req-b54f7baf-59b0-4c6b-9912-8ff68c41ea55 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 955.351056] env[62965]: DEBUG nova.virt.hardware [None req-b54f7baf-59b0-4c6b-9912-8ff68c41ea55 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 955.351210] env[62965]: DEBUG nova.virt.hardware [None req-b54f7baf-59b0-4c6b-9912-8ff68c41ea55 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 955.351430] env[62965]: DEBUG nova.virt.hardware [None req-b54f7baf-59b0-4c6b-9912-8ff68c41ea55 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 955.351594] env[62965]: DEBUG nova.virt.hardware [None req-b54f7baf-59b0-4c6b-9912-8ff68c41ea55 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 955.351765] env[62965]: DEBUG nova.virt.hardware [None req-b54f7baf-59b0-4c6b-9912-8ff68c41ea55 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 955.351925] env[62965]: DEBUG nova.virt.hardware [None req-b54f7baf-59b0-4c6b-9912-8ff68c41ea55 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 955.352109] env[62965]: DEBUG nova.virt.hardware [None req-b54f7baf-59b0-4c6b-9912-8ff68c41ea55 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 955.358246] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b54f7baf-59b0-4c6b-9912-8ff68c41ea55 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Reconfiguring VM to attach interface {{(pid=62965) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 955.358844] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-14a651ee-99f0-4eff-9240-a36e75ac3d6e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.378373] env[62965]: DEBUG oslo_vmware.api [None req-b54f7baf-59b0-4c6b-9912-8ff68c41ea55 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Waiting for the task: (returnval){ [ 955.378373] env[62965]: value = "task-1390134" [ 955.378373] env[62965]: _type = "Task" [ 955.378373] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.387440] env[62965]: DEBUG oslo_vmware.api [None req-b54f7baf-59b0-4c6b-9912-8ff68c41ea55 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390134, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.441969] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.843s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.444403] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.851s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.444635] env[62965]: DEBUG nova.objects.instance [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lazy-loading 'pci_requests' on Instance uuid b59d6023-a047-4b32-bd18-1c7cc05f67d7 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 955.469124] env[62965]: DEBUG oslo_concurrency.lockutils [None req-17cb7e4a-8c99-4681-89b1-eaa6f3d59e3b tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "fbbcbe60-92f3-497d-b190-520a2547b3a9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.469422] env[62965]: DEBUG oslo_concurrency.lockutils [None req-17cb7e4a-8c99-4681-89b1-eaa6f3d59e3b tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "fbbcbe60-92f3-497d-b190-520a2547b3a9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.469637] env[62965]: DEBUG oslo_concurrency.lockutils [None req-17cb7e4a-8c99-4681-89b1-eaa6f3d59e3b tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "fbbcbe60-92f3-497d-b190-520a2547b3a9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.469971] env[62965]: DEBUG oslo_concurrency.lockutils [None req-17cb7e4a-8c99-4681-89b1-eaa6f3d59e3b tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "fbbcbe60-92f3-497d-b190-520a2547b3a9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.470688] env[62965]: DEBUG oslo_concurrency.lockutils [None req-17cb7e4a-8c99-4681-89b1-eaa6f3d59e3b tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "fbbcbe60-92f3-497d-b190-520a2547b3a9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.473078] env[62965]: INFO nova.scheduler.client.report [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Deleted allocations for instance 0eb943ae-462c-40dd-a870-7a884ea2db74 [ 955.477102] env[62965]: INFO nova.compute.manager [None req-17cb7e4a-8c99-4681-89b1-eaa6f3d59e3b tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Terminating instance [ 955.605139] env[62965]: DEBUG oslo_vmware.api [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52cccfd1-8eef-95de-cbce-413263a0eb0a, 'name': SearchDatastore_Task, 'duration_secs': 0.009437} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.605139] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f861c8e6-ec1f-4e30-89b3-a7445e7133ba {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.610201] env[62965]: DEBUG oslo_vmware.api [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Waiting for the task: (returnval){ [ 955.610201] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]527314cd-e5b0-810a-38db-c3301362b877" [ 955.610201] env[62965]: _type = "Task" [ 955.610201] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.615878] env[62965]: DEBUG nova.network.neutron [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Updating instance_info_cache with network_info: [{"id": "6ffa156f-6a68-424e-b324-660892b9ad22", "address": "fa:16:3e:ed:d8:a1", "network": {"id": "e9a6ebdd-0479-45af-b947-5d35ae182c87", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-385052062-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89fd5275b2064288b44e79365f6f0271", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ba07329-1d3e-4ba8-8774-d029262318c4", "external-id": "nsx-vlan-transportzone-534", "segmentation_id": 534, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ffa156f-6a", "ovs_interfaceid": "6ffa156f-6a68-424e-b324-660892b9ad22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.618413] env[62965]: DEBUG nova.compute.manager [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 955.627545] env[62965]: DEBUG oslo_vmware.api [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]527314cd-e5b0-810a-38db-c3301362b877, 'name': SearchDatastore_Task, 'duration_secs': 0.00982} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.628585] env[62965]: DEBUG oslo_concurrency.lockutils [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.628851] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] a63b301d-8d52-4fec-951f-a15cda9449ac/a63b301d-8d52-4fec-951f-a15cda9449ac.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 955.629132] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-90c424f9-6e45-49ba-a807-e3df2fd70595 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.637787] env[62965]: DEBUG oslo_vmware.api [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Waiting for the task: (returnval){ [ 955.637787] env[62965]: value = "task-1390135" [ 955.637787] env[62965]: _type = "Task" [ 955.637787] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.643242] env[62965]: DEBUG nova.virt.hardware [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 955.643482] env[62965]: DEBUG nova.virt.hardware [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 955.643642] env[62965]: DEBUG nova.virt.hardware [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 955.643824] env[62965]: DEBUG nova.virt.hardware [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 955.643971] env[62965]: DEBUG nova.virt.hardware [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 955.644136] env[62965]: DEBUG nova.virt.hardware [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 955.644347] env[62965]: DEBUG nova.virt.hardware [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 955.644511] env[62965]: DEBUG nova.virt.hardware [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 955.644681] env[62965]: DEBUG nova.virt.hardware [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 955.644854] env[62965]: DEBUG nova.virt.hardware [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 955.645144] env[62965]: DEBUG nova.virt.hardware [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 955.646030] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e721d51b-2993-4af5-93a8-d00a43017161 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.656072] env[62965]: DEBUG oslo_vmware.api [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390135, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.663749] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d0b22d9-b69e-490a-8c23-522699e31312 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.667746] env[62965]: DEBUG oslo_vmware.api [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390133, 'name': PowerOffVM_Task, 'duration_secs': 0.197637} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.668640] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 955.668824] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 955.669414] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d59937b2-4db4-4abe-9342-2d291113987e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.749613] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 955.749833] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Deleting contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 955.750053] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Deleting the datastore file [datastore1] 97954de1-7b4b-4615-8d0f-0e8580a515ed {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 955.750741] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-058a653b-ff2b-42a7-b58e-43a86c17a60e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.757581] env[62965]: DEBUG oslo_vmware.api [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for the task: (returnval){ [ 955.757581] env[62965]: value = "task-1390137" [ 955.757581] env[62965]: _type = "Task" [ 955.757581] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.766525] env[62965]: DEBUG oslo_vmware.api [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390137, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.889986] env[62965]: DEBUG oslo_vmware.api [None req-b54f7baf-59b0-4c6b-9912-8ff68c41ea55 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390134, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.949464] env[62965]: DEBUG nova.objects.instance [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lazy-loading 'numa_topology' on Instance uuid b59d6023-a047-4b32-bd18-1c7cc05f67d7 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 955.985130] env[62965]: DEBUG nova.compute.manager [None req-17cb7e4a-8c99-4681-89b1-eaa6f3d59e3b tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 955.985951] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-17cb7e4a-8c99-4681-89b1-eaa6f3d59e3b tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 955.988839] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2364ee2b-b05e-411b-b973-28beae7d2fc9 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "0eb943ae-462c-40dd-a870-7a884ea2db74" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.783s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.991544] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43d00c3e-b515-4d33-8bd6-44a2b1e7c6c9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.002629] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-17cb7e4a-8c99-4681-89b1-eaa6f3d59e3b tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 956.002946] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a2678aa1-0088-4959-a45a-b0ccbc039b66 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.013374] env[62965]: DEBUG oslo_vmware.api [None req-17cb7e4a-8c99-4681-89b1-eaa6f3d59e3b tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 956.013374] env[62965]: value = "task-1390138" [ 956.013374] env[62965]: _type = "Task" [ 956.013374] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.025413] env[62965]: DEBUG oslo_vmware.api [None req-17cb7e4a-8c99-4681-89b1-eaa6f3d59e3b tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390138, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.082644] env[62965]: DEBUG nova.network.neutron [req-1ecdf3f9-afc6-4d9f-b339-a014796be0e1 req-99ed7cee-0cc6-425d-96e5-c57bab759e71 service nova] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Updated VIF entry in instance network info cache for port 9c46a8bf-6b23-45d3-95a4-f0c751895c71. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 956.082881] env[62965]: DEBUG nova.network.neutron [req-1ecdf3f9-afc6-4d9f-b339-a014796be0e1 req-99ed7cee-0cc6-425d-96e5-c57bab759e71 service nova] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Updating instance_info_cache with network_info: [{"id": "8fea9d40-d98f-4328-ad3e-ae708a849908", "address": "fa:16:3e:c3:8c:35", "network": {"id": "a9699d8b-5e76-4666-a236-33ab8f22345e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1329540850-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1952dedf340f4b7ba0d57af6bba9a749", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a316376e-2ef0-4b1e-b40c-10321ebd7e1a", "external-id": "nsx-vlan-transportzone-942", "segmentation_id": 942, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8fea9d40-d9", "ovs_interfaceid": "8fea9d40-d98f-4328-ad3e-ae708a849908", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "9c46a8bf-6b23-45d3-95a4-f0c751895c71", "address": "fa:16:3e:36:f8:d3", "network": {"id": "a9699d8b-5e76-4666-a236-33ab8f22345e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1329540850-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1952dedf340f4b7ba0d57af6bba9a749", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a316376e-2ef0-4b1e-b40c-10321ebd7e1a", "external-id": "nsx-vlan-transportzone-942", "segmentation_id": 942, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c46a8bf-6b", "ovs_interfaceid": "9c46a8bf-6b23-45d3-95a4-f0c751895c71", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.118716] env[62965]: DEBUG oslo_concurrency.lockutils [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Releasing lock "refresh_cache-a0e9cc1a-bc95-4a44-99c5-aabf85a373f4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.125717] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "464a3a31-06bf-4662-a5b1-d185d9a64f0b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.125968] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "464a3a31-06bf-4662-a5b1-d185d9a64f0b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.149335] env[62965]: DEBUG oslo_vmware.api [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390135, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.270900] env[62965]: DEBUG nova.network.neutron [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Successfully updated port: 6b5a028e-baa2-4d3d-a9ca-ade1c685e4b9 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 956.276940] env[62965]: DEBUG oslo_vmware.api [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390137, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.391553] env[62965]: DEBUG oslo_vmware.api [None req-b54f7baf-59b0-4c6b-9912-8ff68c41ea55 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390134, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.404601] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 956.405531] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c11ff8fa-2868-4935-92e0-da78cbad5f84 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.414540] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 956.414829] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-da226d32-ad99-4c5a-838b-ba14b5452e5e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.447481] env[62965]: DEBUG nova.compute.manager [req-de9132a2-7e8d-4423-b388-eaeec1816ac4 req-cde830ca-a5b7-447a-b34e-a42c91cd86da service nova] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Received event network-vif-plugged-6b5a028e-baa2-4d3d-a9ca-ade1c685e4b9 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 956.447893] env[62965]: DEBUG oslo_concurrency.lockutils [req-de9132a2-7e8d-4423-b388-eaeec1816ac4 req-cde830ca-a5b7-447a-b34e-a42c91cd86da service nova] Acquiring lock "08f7b05f-3fb5-48ac-a68a-1d6fd804a622-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.448362] env[62965]: DEBUG oslo_concurrency.lockutils [req-de9132a2-7e8d-4423-b388-eaeec1816ac4 req-cde830ca-a5b7-447a-b34e-a42c91cd86da service nova] Lock "08f7b05f-3fb5-48ac-a68a-1d6fd804a622-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.448558] env[62965]: DEBUG oslo_concurrency.lockutils [req-de9132a2-7e8d-4423-b388-eaeec1816ac4 req-cde830ca-a5b7-447a-b34e-a42c91cd86da service nova] Lock "08f7b05f-3fb5-48ac-a68a-1d6fd804a622-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.448732] env[62965]: DEBUG nova.compute.manager [req-de9132a2-7e8d-4423-b388-eaeec1816ac4 req-cde830ca-a5b7-447a-b34e-a42c91cd86da service nova] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] No waiting events found dispatching network-vif-plugged-6b5a028e-baa2-4d3d-a9ca-ade1c685e4b9 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 956.449087] env[62965]: WARNING nova.compute.manager [req-de9132a2-7e8d-4423-b388-eaeec1816ac4 req-cde830ca-a5b7-447a-b34e-a42c91cd86da service nova] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Received unexpected event network-vif-plugged-6b5a028e-baa2-4d3d-a9ca-ade1c685e4b9 for instance with vm_state building and task_state spawning. [ 956.452820] env[62965]: INFO nova.compute.claims [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 956.527045] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 956.527238] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Deleting contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 956.527361] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Deleting the datastore file [datastore1] a0e9cc1a-bc95-4a44-99c5-aabf85a373f4 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 956.528587] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fcfec6ab-db52-4711-810b-8331c019ac29 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.535973] env[62965]: DEBUG oslo_vmware.api [None req-17cb7e4a-8c99-4681-89b1-eaa6f3d59e3b tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390138, 'name': PowerOffVM_Task, 'duration_secs': 0.38028} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.535973] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-17cb7e4a-8c99-4681-89b1-eaa6f3d59e3b tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 956.535973] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-17cb7e4a-8c99-4681-89b1-eaa6f3d59e3b tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 956.535973] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a4d22d8b-2dd9-4298-bcc1-e44ceece71d3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.539511] env[62965]: DEBUG oslo_vmware.api [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 956.539511] env[62965]: value = "task-1390140" [ 956.539511] env[62965]: _type = "Task" [ 956.539511] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.549827] env[62965]: DEBUG oslo_vmware.api [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390140, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.585779] env[62965]: DEBUG oslo_concurrency.lockutils [req-1ecdf3f9-afc6-4d9f-b339-a014796be0e1 req-99ed7cee-0cc6-425d-96e5-c57bab759e71 service nova] Releasing lock "refresh_cache-00f24976-2f0e-4a2f-8f2e-012b0fce479f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.629019] env[62965]: DEBUG nova.compute.manager [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 956.633429] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-17cb7e4a-8c99-4681-89b1-eaa6f3d59e3b tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 956.633429] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-17cb7e4a-8c99-4681-89b1-eaa6f3d59e3b tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Deleting contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 956.633429] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-17cb7e4a-8c99-4681-89b1-eaa6f3d59e3b tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Deleting the datastore file [datastore2] fbbcbe60-92f3-497d-b190-520a2547b3a9 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 956.633877] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b3c8ec4d-2a91-4518-b14d-8797a0a22282 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.642602] env[62965]: DEBUG oslo_vmware.api [None req-17cb7e4a-8c99-4681-89b1-eaa6f3d59e3b tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 956.642602] env[62965]: value = "task-1390142" [ 956.642602] env[62965]: _type = "Task" [ 956.642602] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.650016] env[62965]: DEBUG oslo_vmware.api [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390135, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.522896} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.650712] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] a63b301d-8d52-4fec-951f-a15cda9449ac/a63b301d-8d52-4fec-951f-a15cda9449ac.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 956.650961] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 956.651445] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-25b641b0-b169-4b3c-97cb-7704e2b6df09 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.656669] env[62965]: DEBUG oslo_vmware.api [None req-17cb7e4a-8c99-4681-89b1-eaa6f3d59e3b tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390142, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.662646] env[62965]: DEBUG oslo_vmware.api [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Waiting for the task: (returnval){ [ 956.662646] env[62965]: value = "task-1390143" [ 956.662646] env[62965]: _type = "Task" [ 956.662646] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.672939] env[62965]: DEBUG oslo_vmware.api [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390143, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.769032] env[62965]: DEBUG oslo_vmware.api [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390137, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.770656] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "refresh_cache-08f7b05f-3fb5-48ac-a68a-1d6fd804a622" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.770789] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquired lock "refresh_cache-08f7b05f-3fb5-48ac-a68a-1d6fd804a622" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.770932] env[62965]: DEBUG nova.network.neutron [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 956.889832] env[62965]: DEBUG oslo_vmware.api [None req-b54f7baf-59b0-4c6b-9912-8ff68c41ea55 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390134, 'name': ReconfigVM_Task, 'duration_secs': 1.178415} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.890403] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b54f7baf-59b0-4c6b-9912-8ff68c41ea55 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Releasing lock "00f24976-2f0e-4a2f-8f2e-012b0fce479f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.890629] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b54f7baf-59b0-4c6b-9912-8ff68c41ea55 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Reconfigured VM to attach interface {{(pid=62965) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 957.050051] env[62965]: DEBUG oslo_vmware.api [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390140, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.070428] env[62965]: DEBUG nova.compute.manager [req-2940d6d2-2b2a-42e4-b056-8f692213ee5a req-90e60a0e-1f36-411f-9e84-7044f57a91ff service nova] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Received event network-vif-unplugged-6ffa156f-6a68-424e-b324-660892b9ad22 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 957.070728] env[62965]: DEBUG oslo_concurrency.lockutils [req-2940d6d2-2b2a-42e4-b056-8f692213ee5a req-90e60a0e-1f36-411f-9e84-7044f57a91ff service nova] Acquiring lock "a0e9cc1a-bc95-4a44-99c5-aabf85a373f4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.070908] env[62965]: DEBUG oslo_concurrency.lockutils [req-2940d6d2-2b2a-42e4-b056-8f692213ee5a req-90e60a0e-1f36-411f-9e84-7044f57a91ff service nova] Lock "a0e9cc1a-bc95-4a44-99c5-aabf85a373f4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.071186] env[62965]: DEBUG oslo_concurrency.lockutils [req-2940d6d2-2b2a-42e4-b056-8f692213ee5a req-90e60a0e-1f36-411f-9e84-7044f57a91ff service nova] Lock "a0e9cc1a-bc95-4a44-99c5-aabf85a373f4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.071439] env[62965]: DEBUG nova.compute.manager [req-2940d6d2-2b2a-42e4-b056-8f692213ee5a req-90e60a0e-1f36-411f-9e84-7044f57a91ff service nova] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] No waiting events found dispatching network-vif-unplugged-6ffa156f-6a68-424e-b324-660892b9ad22 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 957.071649] env[62965]: WARNING nova.compute.manager [req-2940d6d2-2b2a-42e4-b056-8f692213ee5a req-90e60a0e-1f36-411f-9e84-7044f57a91ff service nova] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Received unexpected event network-vif-unplugged-6ffa156f-6a68-424e-b324-660892b9ad22 for instance with vm_state shelved and task_state shelving_offloading. [ 957.071825] env[62965]: DEBUG nova.compute.manager [req-2940d6d2-2b2a-42e4-b056-8f692213ee5a req-90e60a0e-1f36-411f-9e84-7044f57a91ff service nova] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Received event network-changed-6ffa156f-6a68-424e-b324-660892b9ad22 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 957.071987] env[62965]: DEBUG nova.compute.manager [req-2940d6d2-2b2a-42e4-b056-8f692213ee5a req-90e60a0e-1f36-411f-9e84-7044f57a91ff service nova] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Refreshing instance network info cache due to event network-changed-6ffa156f-6a68-424e-b324-660892b9ad22. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 957.072201] env[62965]: DEBUG oslo_concurrency.lockutils [req-2940d6d2-2b2a-42e4-b056-8f692213ee5a req-90e60a0e-1f36-411f-9e84-7044f57a91ff service nova] Acquiring lock "refresh_cache-a0e9cc1a-bc95-4a44-99c5-aabf85a373f4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.072347] env[62965]: DEBUG oslo_concurrency.lockutils [req-2940d6d2-2b2a-42e4-b056-8f692213ee5a req-90e60a0e-1f36-411f-9e84-7044f57a91ff service nova] Acquired lock "refresh_cache-a0e9cc1a-bc95-4a44-99c5-aabf85a373f4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.072559] env[62965]: DEBUG nova.network.neutron [req-2940d6d2-2b2a-42e4-b056-8f692213ee5a req-90e60a0e-1f36-411f-9e84-7044f57a91ff service nova] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Refreshing network info cache for port 6ffa156f-6a68-424e-b324-660892b9ad22 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 957.148977] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.152919] env[62965]: DEBUG oslo_vmware.api [None req-17cb7e4a-8c99-4681-89b1-eaa6f3d59e3b tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390142, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150965} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.153177] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-17cb7e4a-8c99-4681-89b1-eaa6f3d59e3b tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 957.153366] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-17cb7e4a-8c99-4681-89b1-eaa6f3d59e3b tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Deleted contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 957.153549] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-17cb7e4a-8c99-4681-89b1-eaa6f3d59e3b tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 957.153721] env[62965]: INFO nova.compute.manager [None req-17cb7e4a-8c99-4681-89b1-eaa6f3d59e3b tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Took 1.17 seconds to destroy the instance on the hypervisor. [ 957.153964] env[62965]: DEBUG oslo.service.loopingcall [None req-17cb7e4a-8c99-4681-89b1-eaa6f3d59e3b tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 957.154173] env[62965]: DEBUG nova.compute.manager [-] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 957.154269] env[62965]: DEBUG nova.network.neutron [-] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 957.172395] env[62965]: DEBUG oslo_vmware.api [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390143, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.099397} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.172646] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 957.173432] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63e4402f-8f5e-489c-9547-42c5e3569e71 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.197225] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] a63b301d-8d52-4fec-951f-a15cda9449ac/a63b301d-8d52-4fec-951f-a15cda9449ac.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 957.197549] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f47f4fc3-e32b-405c-9647-5586acec7dd0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.220983] env[62965]: DEBUG oslo_vmware.api [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Waiting for the task: (returnval){ [ 957.220983] env[62965]: value = "task-1390144" [ 957.220983] env[62965]: _type = "Task" [ 957.220983] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.229136] env[62965]: DEBUG oslo_vmware.api [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390144, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.270332] env[62965]: DEBUG oslo_vmware.api [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390137, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.395201] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b54f7baf-59b0-4c6b-9912-8ff68c41ea55 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "interface-00f24976-2f0e-4a2f-8f2e-012b0fce479f-9c46a8bf-6b23-45d3-95a4-f0c751895c71" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.823s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.486849] env[62965]: DEBUG nova.network.neutron [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 957.554705] env[62965]: DEBUG oslo_vmware.api [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390140, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.638724] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e3ef927-6bc3-4733-a138-88f1be7d8d31 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.650270] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e154d662-0f88-4e82-9a0c-a073e7ce3158 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.683135] env[62965]: DEBUG nova.network.neutron [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Updating instance_info_cache with network_info: [{"id": "6b5a028e-baa2-4d3d-a9ca-ade1c685e4b9", "address": "fa:16:3e:3b:4e:09", "network": {"id": "a0a5e605-d580-4b7b-b3c0-5c7395bd5a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1679618017-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ed71548c085499981fa1b102d660368", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b5a028e-ba", "ovs_interfaceid": "6b5a028e-baa2-4d3d-a9ca-ade1c685e4b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.684856] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f0b47e4-568e-45d4-98e4-d8c5605ff579 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.694353] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e6fb932-ee23-4b49-b38f-6fa3b54b9805 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.709941] env[62965]: DEBUG nova.compute.provider_tree [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 957.733604] env[62965]: DEBUG oslo_vmware.api [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390144, 'name': ReconfigVM_Task, 'duration_secs': 0.337669} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.733887] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Reconfigured VM instance instance-00000058 to attach disk [datastore2] a63b301d-8d52-4fec-951f-a15cda9449ac/a63b301d-8d52-4fec-951f-a15cda9449ac.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 957.734510] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-94b28763-e1c0-4ff5-a471-1d78c6e96ce9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.744575] env[62965]: DEBUG oslo_vmware.api [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Waiting for the task: (returnval){ [ 957.744575] env[62965]: value = "task-1390145" [ 957.744575] env[62965]: _type = "Task" [ 957.744575] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.754521] env[62965]: DEBUG oslo_vmware.api [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390145, 'name': Rename_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.770259] env[62965]: DEBUG oslo_vmware.api [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390137, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.810197] env[62965]: DEBUG nova.network.neutron [req-2940d6d2-2b2a-42e4-b056-8f692213ee5a req-90e60a0e-1f36-411f-9e84-7044f57a91ff service nova] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Updated VIF entry in instance network info cache for port 6ffa156f-6a68-424e-b324-660892b9ad22. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 957.810604] env[62965]: DEBUG nova.network.neutron [req-2940d6d2-2b2a-42e4-b056-8f692213ee5a req-90e60a0e-1f36-411f-9e84-7044f57a91ff service nova] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Updating instance_info_cache with network_info: [{"id": "6ffa156f-6a68-424e-b324-660892b9ad22", "address": "fa:16:3e:ed:d8:a1", "network": {"id": "e9a6ebdd-0479-45af-b947-5d35ae182c87", "bridge": null, "label": "tempest-ServersNegativeTestJSON-385052062-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89fd5275b2064288b44e79365f6f0271", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap6ffa156f-6a", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.902557] env[62965]: DEBUG nova.network.neutron [-] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.051027] env[62965]: DEBUG oslo_vmware.api [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390140, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.189238] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Releasing lock "refresh_cache-08f7b05f-3fb5-48ac-a68a-1d6fd804a622" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.189566] env[62965]: DEBUG nova.compute.manager [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Instance network_info: |[{"id": "6b5a028e-baa2-4d3d-a9ca-ade1c685e4b9", "address": "fa:16:3e:3b:4e:09", "network": {"id": "a0a5e605-d580-4b7b-b3c0-5c7395bd5a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1679618017-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ed71548c085499981fa1b102d660368", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b5a028e-ba", "ovs_interfaceid": "6b5a028e-baa2-4d3d-a9ca-ade1c685e4b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 958.190014] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3b:4e:09', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0cd5d325-3053-407e-a4ee-f627e82a23f9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6b5a028e-baa2-4d3d-a9ca-ade1c685e4b9', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 958.199221] env[62965]: DEBUG oslo.service.loopingcall [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 958.199439] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 958.199709] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-430f1c6d-3318-480c-900a-b181bc79e855 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.216738] env[62965]: DEBUG nova.scheduler.client.report [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 958.226754] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 958.226754] env[62965]: value = "task-1390146" [ 958.226754] env[62965]: _type = "Task" [ 958.226754] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.238704] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390146, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.255842] env[62965]: DEBUG oslo_vmware.api [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390145, 'name': Rename_Task, 'duration_secs': 0.217887} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.256161] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 958.256427] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-70f1f2dd-ffc2-4fbb-9bc1-636ac074e8ce {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.267457] env[62965]: DEBUG oslo_vmware.api [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Waiting for the task: (returnval){ [ 958.267457] env[62965]: value = "task-1390147" [ 958.267457] env[62965]: _type = "Task" [ 958.267457] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.271257] env[62965]: DEBUG oslo_vmware.api [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390137, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.280636] env[62965]: DEBUG oslo_vmware.api [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390147, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.315029] env[62965]: DEBUG oslo_concurrency.lockutils [req-2940d6d2-2b2a-42e4-b056-8f692213ee5a req-90e60a0e-1f36-411f-9e84-7044f57a91ff service nova] Releasing lock "refresh_cache-a0e9cc1a-bc95-4a44-99c5-aabf85a373f4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.326725] env[62965]: DEBUG oslo_concurrency.lockutils [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "51babf43-b7b0-4731-a989-6977874fb801" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.327064] env[62965]: DEBUG oslo_concurrency.lockutils [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "51babf43-b7b0-4731-a989-6977874fb801" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.406442] env[62965]: INFO nova.compute.manager [-] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Took 1.25 seconds to deallocate network for instance. [ 958.477875] env[62965]: DEBUG nova.compute.manager [req-b6a9142a-0b31-4286-bc6a-04da9cd999ab req-5b0e4412-5adc-47a9-a668-57ca9088244d service nova] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Received event network-changed-6b5a028e-baa2-4d3d-a9ca-ade1c685e4b9 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 958.478179] env[62965]: DEBUG nova.compute.manager [req-b6a9142a-0b31-4286-bc6a-04da9cd999ab req-5b0e4412-5adc-47a9-a668-57ca9088244d service nova] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Refreshing instance network info cache due to event network-changed-6b5a028e-baa2-4d3d-a9ca-ade1c685e4b9. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 958.478506] env[62965]: DEBUG oslo_concurrency.lockutils [req-b6a9142a-0b31-4286-bc6a-04da9cd999ab req-5b0e4412-5adc-47a9-a668-57ca9088244d service nova] Acquiring lock "refresh_cache-08f7b05f-3fb5-48ac-a68a-1d6fd804a622" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.478708] env[62965]: DEBUG oslo_concurrency.lockutils [req-b6a9142a-0b31-4286-bc6a-04da9cd999ab req-5b0e4412-5adc-47a9-a668-57ca9088244d service nova] Acquired lock "refresh_cache-08f7b05f-3fb5-48ac-a68a-1d6fd804a622" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.478890] env[62965]: DEBUG nova.network.neutron [req-b6a9142a-0b31-4286-bc6a-04da9cd999ab req-5b0e4412-5adc-47a9-a668-57ca9088244d service nova] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Refreshing network info cache for port 6b5a028e-baa2-4d3d-a9ca-ade1c685e4b9 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 958.552383] env[62965]: DEBUG oslo_vmware.api [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390140, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.722180] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.278s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.724266] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 5.040s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.737592] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390146, 'name': CreateVM_Task, 'duration_secs': 0.386504} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.737804] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 958.738480] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.738892] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.739041] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 958.739261] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d7cd00de-758c-47df-9068-5c0eb04926ad {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.746434] env[62965]: DEBUG oslo_vmware.api [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 958.746434] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52a0b6ee-09c5-87eb-5eee-676baf5852d4" [ 958.746434] env[62965]: _type = "Task" [ 958.746434] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.756458] env[62965]: DEBUG oslo_vmware.api [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52a0b6ee-09c5-87eb-5eee-676baf5852d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.757377] env[62965]: INFO nova.network.neutron [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Updating port d2d44211-6856-42ed-8f9f-3b1a58b31349 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 958.771915] env[62965]: DEBUG oslo_vmware.api [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Task: {'id': task-1390137, 'name': DeleteDatastoreFile_Task, 'duration_secs': 2.720565} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.776218] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 958.776419] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Deleted contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 958.776595] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 958.776771] env[62965]: INFO nova.compute.manager [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Took 3.65 seconds to destroy the instance on the hypervisor. [ 958.777111] env[62965]: DEBUG oslo.service.loopingcall [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 958.777527] env[62965]: DEBUG nova.compute.manager [-] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 958.777629] env[62965]: DEBUG nova.network.neutron [-] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 958.785884] env[62965]: DEBUG oslo_vmware.api [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390147, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.833011] env[62965]: DEBUG nova.compute.manager [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 958.867025] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7a739b05-b8d3-437c-bbd4-3851e7b35db0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "interface-00f24976-2f0e-4a2f-8f2e-012b0fce479f-9c46a8bf-6b23-45d3-95a4-f0c751895c71" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.867346] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7a739b05-b8d3-437c-bbd4-3851e7b35db0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "interface-00f24976-2f0e-4a2f-8f2e-012b0fce479f-9c46a8bf-6b23-45d3-95a4-f0c751895c71" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.914197] env[62965]: DEBUG oslo_concurrency.lockutils [None req-17cb7e4a-8c99-4681-89b1-eaa6f3d59e3b tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.052320] env[62965]: DEBUG oslo_vmware.api [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390140, 'name': DeleteDatastoreFile_Task, 'duration_secs': 2.105205} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.052481] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 959.052665] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Deleted contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 959.052841] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 959.093019] env[62965]: INFO nova.scheduler.client.report [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Deleted allocations for instance a0e9cc1a-bc95-4a44-99c5-aabf85a373f4 [ 959.149466] env[62965]: DEBUG nova.compute.manager [req-0f83550a-a00d-4e3f-a40f-224ffe89360e req-80a6f486-3838-448c-a983-9c1afef4da88 service nova] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Received event network-vif-deleted-3edb59ed-a2b8-4d31-b975-9c9d64c3a22b {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 959.225489] env[62965]: DEBUG nova.network.neutron [req-b6a9142a-0b31-4286-bc6a-04da9cd999ab req-5b0e4412-5adc-47a9-a668-57ca9088244d service nova] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Updated VIF entry in instance network info cache for port 6b5a028e-baa2-4d3d-a9ca-ade1c685e4b9. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 959.226022] env[62965]: DEBUG nova.network.neutron [req-b6a9142a-0b31-4286-bc6a-04da9cd999ab req-5b0e4412-5adc-47a9-a668-57ca9088244d service nova] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Updating instance_info_cache with network_info: [{"id": "6b5a028e-baa2-4d3d-a9ca-ade1c685e4b9", "address": "fa:16:3e:3b:4e:09", "network": {"id": "a0a5e605-d580-4b7b-b3c0-5c7395bd5a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1679618017-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ed71548c085499981fa1b102d660368", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6b5a028e-ba", "ovs_interfaceid": "6b5a028e-baa2-4d3d-a9ca-ade1c685e4b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.257472] env[62965]: DEBUG oslo_vmware.api [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52a0b6ee-09c5-87eb-5eee-676baf5852d4, 'name': SearchDatastore_Task, 'duration_secs': 0.01256} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.257734] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.257971] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 959.258279] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.258439] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.258621] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 959.258885] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f5e8b76b-d590-4d5e-8fd5-071df94a3191 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.268421] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 959.268708] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 959.269437] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5246bed4-b6c0-4b35-ad56-43c67abccee7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.279381] env[62965]: DEBUG oslo_vmware.api [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 959.279381] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52649951-cf47-e519-7442-f8d694a6f80c" [ 959.279381] env[62965]: _type = "Task" [ 959.279381] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.282198] env[62965]: DEBUG oslo_vmware.api [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390147, 'name': PowerOnVM_Task, 'duration_secs': 0.524043} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.286981] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 959.287217] env[62965]: INFO nova.compute.manager [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Took 8.43 seconds to spawn the instance on the hypervisor. [ 959.287404] env[62965]: DEBUG nova.compute.manager [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 959.288161] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28a4dd16-6f78-4e35-8880-c2e2fb29b879 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.298535] env[62965]: DEBUG oslo_vmware.api [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52649951-cf47-e519-7442-f8d694a6f80c, 'name': SearchDatastore_Task, 'duration_secs': 0.009353} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.301916] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f067fccb-1bd8-4e13-8d24-cee9d8609da2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.308286] env[62965]: DEBUG oslo_vmware.api [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 959.308286] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]528468df-0259-0fa2-a252-b519dea82046" [ 959.308286] env[62965]: _type = "Task" [ 959.308286] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.316302] env[62965]: DEBUG oslo_vmware.api [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]528468df-0259-0fa2-a252-b519dea82046, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.353319] env[62965]: DEBUG oslo_concurrency.lockutils [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.370534] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7a739b05-b8d3-437c-bbd4-3851e7b35db0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "00f24976-2f0e-4a2f-8f2e-012b0fce479f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.370715] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7a739b05-b8d3-437c-bbd4-3851e7b35db0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquired lock "00f24976-2f0e-4a2f-8f2e-012b0fce479f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.371676] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9124c9ef-ce71-41e2-8e20-a208ef25a7d2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.391105] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-633d3585-7637-4309-9d1d-2541a493c136 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.419017] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7a739b05-b8d3-437c-bbd4-3851e7b35db0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Reconfiguring VM to detach interface {{(pid=62965) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 959.419372] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b6467e60-8b25-4c2e-970d-675fc1ce8bce {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.439208] env[62965]: DEBUG oslo_vmware.api [None req-7a739b05-b8d3-437c-bbd4-3851e7b35db0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Waiting for the task: (returnval){ [ 959.439208] env[62965]: value = "task-1390148" [ 959.439208] env[62965]: _type = "Task" [ 959.439208] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.448011] env[62965]: DEBUG oslo_vmware.api [None req-7a739b05-b8d3-437c-bbd4-3851e7b35db0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390148, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.596343] env[62965]: DEBUG oslo_concurrency.lockutils [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.602853] env[62965]: DEBUG nova.network.neutron [-] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.736379] env[62965]: DEBUG oslo_concurrency.lockutils [req-b6a9142a-0b31-4286-bc6a-04da9cd999ab req-5b0e4412-5adc-47a9-a668-57ca9088244d service nova] Releasing lock "refresh_cache-08f7b05f-3fb5-48ac-a68a-1d6fd804a622" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.754247] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance c2448b76-1553-4b68-a731-0bb0967f4c1d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 959.754399] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 924c5cdc-0450-4ced-b920-50e2d5060fd2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 959.754517] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance bd36601b-5a70-4a29-8ef1-d1e925f41de7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 959.754631] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance d038608c-9264-4fd5-b9ba-0653c98ec0e4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 959.754743] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 00f24976-2f0e-4a2f-8f2e-012b0fce479f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 959.754889] env[62965]: WARNING nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance fbbcbe60-92f3-497d-b190-520a2547b3a9 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 959.755010] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 97954de1-7b4b-4615-8d0f-0e8580a515ed actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 959.755145] env[62965]: WARNING nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 3223b1ed-abc1-439c-bcad-6317eb204e45 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 959.755354] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance a63b301d-8d52-4fec-951f-a15cda9449ac actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 959.755486] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance b59d6023-a047-4b32-bd18-1c7cc05f67d7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 959.755597] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 08f7b05f-3fb5-48ac-a68a-1d6fd804a622 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 959.812534] env[62965]: INFO nova.compute.manager [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Took 19.12 seconds to build instance. [ 959.819433] env[62965]: DEBUG oslo_vmware.api [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]528468df-0259-0fa2-a252-b519dea82046, 'name': SearchDatastore_Task, 'duration_secs': 0.009135} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.819706] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.819956] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 08f7b05f-3fb5-48ac-a68a-1d6fd804a622/08f7b05f-3fb5-48ac-a68a-1d6fd804a622.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 959.820280] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5fc174c4-a953-478a-8b4e-7ff1bd14fd3e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.827222] env[62965]: DEBUG oslo_vmware.api [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 959.827222] env[62965]: value = "task-1390149" [ 959.827222] env[62965]: _type = "Task" [ 959.827222] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.838144] env[62965]: DEBUG oslo_vmware.api [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390149, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.950445] env[62965]: DEBUG oslo_vmware.api [None req-7a739b05-b8d3-437c-bbd4-3851e7b35db0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390148, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.105383] env[62965]: INFO nova.compute.manager [-] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Took 1.33 seconds to deallocate network for instance. [ 960.258730] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 464a3a31-06bf-4662-a5b1-d185d9a64f0b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 960.315338] env[62965]: DEBUG oslo_concurrency.lockutils [None req-996a82c9-a498-4d0d-af5b-2b93b443b774 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "a63b301d-8d52-4fec-951f-a15cda9449ac" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.631s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.339114] env[62965]: DEBUG oslo_vmware.api [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390149, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.488044} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.339495] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 08f7b05f-3fb5-48ac-a68a-1d6fd804a622/08f7b05f-3fb5-48ac-a68a-1d6fd804a622.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 960.339720] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 960.340336] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ade3da51-9360-4614-a6d8-0de8ba14cd70 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.347434] env[62965]: DEBUG oslo_vmware.api [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 960.347434] env[62965]: value = "task-1390150" [ 960.347434] env[62965]: _type = "Task" [ 960.347434] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.356356] env[62965]: DEBUG oslo_vmware.api [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390150, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.388346] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquiring lock "refresh_cache-b59d6023-a047-4b32-bd18-1c7cc05f67d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.388550] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquired lock "refresh_cache-b59d6023-a047-4b32-bd18-1c7cc05f67d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.388735] env[62965]: DEBUG nova.network.neutron [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 960.449979] env[62965]: DEBUG oslo_vmware.api [None req-7a739b05-b8d3-437c-bbd4-3851e7b35db0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390148, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.522511] env[62965]: DEBUG nova.compute.manager [req-7009d794-15e1-42df-a995-9097e890e4e4 req-df656c65-bcac-482b-83d1-223e2b196079 service nova] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Received event network-vif-deleted-212a2219-fc24-44bb-8340-6ad61740edf0 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 960.612928] env[62965]: DEBUG oslo_concurrency.lockutils [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.763848] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 51babf43-b7b0-4731-a989-6977874fb801 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 960.764619] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Total usable vcpus: 48, total allocated vcpus: 10 {{(pid=62965) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 960.764619] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2432MB phys_disk=200GB used_disk=10GB total_vcpus=48 used_vcpus=10 pci_stats=[] {{(pid=62965) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 960.863681] env[62965]: DEBUG oslo_vmware.api [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390150, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079333} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.863979] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 960.864841] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92c3104f-60b8-4a89-b6a6-9103e94f5489 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.906589] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] 08f7b05f-3fb5-48ac-a68a-1d6fd804a622/08f7b05f-3fb5-48ac-a68a-1d6fd804a622.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 960.914844] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-39f6ee8e-0f70-4c66-96f6-77c0dce39225 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.959392] env[62965]: DEBUG oslo_vmware.api [None req-7a739b05-b8d3-437c-bbd4-3851e7b35db0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390148, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.964715] env[62965]: DEBUG oslo_vmware.api [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 960.964715] env[62965]: value = "task-1390151" [ 960.964715] env[62965]: _type = "Task" [ 960.964715] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.977787] env[62965]: DEBUG oslo_vmware.api [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390151, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.035317] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa4fabb6-aa2c-428f-9d31-fb12fb79b01c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.044611] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36ca5d8b-adeb-452c-bd98-9b1757edf73e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.079187] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77f9af41-82b3-4a2c-92db-de340d155ab4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.087834] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-437f0611-ccc7-4401-b817-d45a7b1f9181 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.101763] env[62965]: DEBUG nova.compute.provider_tree [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 961.176109] env[62965]: DEBUG nova.compute.manager [req-c143305f-b926-49f9-835c-f8f9c80da69c req-4a780641-e2cf-4cba-b5e3-ca1faf90a80b service nova] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Received event network-vif-plugged-d2d44211-6856-42ed-8f9f-3b1a58b31349 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 961.176358] env[62965]: DEBUG oslo_concurrency.lockutils [req-c143305f-b926-49f9-835c-f8f9c80da69c req-4a780641-e2cf-4cba-b5e3-ca1faf90a80b service nova] Acquiring lock "b59d6023-a047-4b32-bd18-1c7cc05f67d7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.176578] env[62965]: DEBUG oslo_concurrency.lockutils [req-c143305f-b926-49f9-835c-f8f9c80da69c req-4a780641-e2cf-4cba-b5e3-ca1faf90a80b service nova] Lock "b59d6023-a047-4b32-bd18-1c7cc05f67d7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.176750] env[62965]: DEBUG oslo_concurrency.lockutils [req-c143305f-b926-49f9-835c-f8f9c80da69c req-4a780641-e2cf-4cba-b5e3-ca1faf90a80b service nova] Lock "b59d6023-a047-4b32-bd18-1c7cc05f67d7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.176941] env[62965]: DEBUG nova.compute.manager [req-c143305f-b926-49f9-835c-f8f9c80da69c req-4a780641-e2cf-4cba-b5e3-ca1faf90a80b service nova] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] No waiting events found dispatching network-vif-plugged-d2d44211-6856-42ed-8f9f-3b1a58b31349 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 961.178095] env[62965]: WARNING nova.compute.manager [req-c143305f-b926-49f9-835c-f8f9c80da69c req-4a780641-e2cf-4cba-b5e3-ca1faf90a80b service nova] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Received unexpected event network-vif-plugged-d2d44211-6856-42ed-8f9f-3b1a58b31349 for instance with vm_state shelved_offloaded and task_state spawning. [ 961.178095] env[62965]: DEBUG nova.compute.manager [req-c143305f-b926-49f9-835c-f8f9c80da69c req-4a780641-e2cf-4cba-b5e3-ca1faf90a80b service nova] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Received event network-changed-d2d44211-6856-42ed-8f9f-3b1a58b31349 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 961.178095] env[62965]: DEBUG nova.compute.manager [req-c143305f-b926-49f9-835c-f8f9c80da69c req-4a780641-e2cf-4cba-b5e3-ca1faf90a80b service nova] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Refreshing instance network info cache due to event network-changed-d2d44211-6856-42ed-8f9f-3b1a58b31349. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 961.178095] env[62965]: DEBUG oslo_concurrency.lockutils [req-c143305f-b926-49f9-835c-f8f9c80da69c req-4a780641-e2cf-4cba-b5e3-ca1faf90a80b service nova] Acquiring lock "refresh_cache-b59d6023-a047-4b32-bd18-1c7cc05f67d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.214421] env[62965]: DEBUG nova.network.neutron [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Updating instance_info_cache with network_info: [{"id": "d2d44211-6856-42ed-8f9f-3b1a58b31349", "address": "fa:16:3e:3b:f0:e9", "network": {"id": "49dd42b0-6397-4fdd-a319-ed920c72a5af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-373172566-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.230", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d91c99b14018469e8998fd1ccc0fab4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2d44211-68", "ovs_interfaceid": "d2d44211-6856-42ed-8f9f-3b1a58b31349", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.456233] env[62965]: DEBUG oslo_vmware.api [None req-7a739b05-b8d3-437c-bbd4-3851e7b35db0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390148, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.474474] env[62965]: DEBUG oslo_vmware.api [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390151, 'name': ReconfigVM_Task, 'duration_secs': 0.293751} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.474801] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Reconfigured VM instance instance-00000059 to attach disk [datastore2] 08f7b05f-3fb5-48ac-a68a-1d6fd804a622/08f7b05f-3fb5-48ac-a68a-1d6fd804a622.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 961.475446] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e93c73ae-ae05-42f3-9c50-c8106eeeec49 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.483359] env[62965]: DEBUG oslo_vmware.api [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 961.483359] env[62965]: value = "task-1390152" [ 961.483359] env[62965]: _type = "Task" [ 961.483359] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.492147] env[62965]: DEBUG oslo_vmware.api [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390152, 'name': Rename_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.624542] env[62965]: ERROR nova.scheduler.client.report [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [req-54de0073-d596-4afc-8e1d-1cc942bede0f] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-54de0073-d596-4afc-8e1d-1cc942bede0f"}]} [ 961.640310] env[62965]: DEBUG nova.scheduler.client.report [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Refreshing inventories for resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 961.654705] env[62965]: DEBUG nova.scheduler.client.report [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Updating ProviderTree inventory for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 961.654890] env[62965]: DEBUG nova.compute.provider_tree [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 961.666531] env[62965]: DEBUG nova.scheduler.client.report [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Refreshing aggregate associations for resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8, aggregates: None {{(pid=62965) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 961.674185] env[62965]: DEBUG oslo_concurrency.lockutils [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquiring lock "a0e9cc1a-bc95-4a44-99c5-aabf85a373f4" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.686035] env[62965]: DEBUG nova.scheduler.client.report [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Refreshing trait associations for resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64 {{(pid=62965) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 961.717126] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Releasing lock "refresh_cache-b59d6023-a047-4b32-bd18-1c7cc05f67d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.719657] env[62965]: DEBUG oslo_concurrency.lockutils [req-c143305f-b926-49f9-835c-f8f9c80da69c req-4a780641-e2cf-4cba-b5e3-ca1faf90a80b service nova] Acquired lock "refresh_cache-b59d6023-a047-4b32-bd18-1c7cc05f67d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.719657] env[62965]: DEBUG nova.network.neutron [req-c143305f-b926-49f9-835c-f8f9c80da69c req-4a780641-e2cf-4cba-b5e3-ca1faf90a80b service nova] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Refreshing network info cache for port d2d44211-6856-42ed-8f9f-3b1a58b31349 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 961.743706] env[62965]: DEBUG nova.virt.hardware [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='0ff8788c5773bdf80e990f3cc62d8847',container_format='bare',created_at=2024-10-31T12:37:27Z,direct_url=,disk_format='vmdk',id=c93dd56b-480e-46d6-9aed-4d5aa28dd1e0,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1700694244-shelved',owner='d91c99b14018469e8998fd1ccc0fab4f',properties=ImageMetaProps,protected=,size=31666176,status='active',tags=,updated_at=2024-10-31T12:37:42Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 961.743955] env[62965]: DEBUG nova.virt.hardware [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 961.744128] env[62965]: DEBUG nova.virt.hardware [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 961.744318] env[62965]: DEBUG nova.virt.hardware [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 961.744464] env[62965]: DEBUG nova.virt.hardware [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 961.744612] env[62965]: DEBUG nova.virt.hardware [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 961.744819] env[62965]: DEBUG nova.virt.hardware [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 961.744986] env[62965]: DEBUG nova.virt.hardware [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 961.745163] env[62965]: DEBUG nova.virt.hardware [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 961.745332] env[62965]: DEBUG nova.virt.hardware [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 961.745711] env[62965]: DEBUG nova.virt.hardware [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 961.746838] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4479bc3-8b8a-4829-b24d-653effe3833e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.758580] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c5135f1-42f3-4e60-b6da-93097de96243 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.773514] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3b:f0:e9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f77ff7a1-209c-4f3f-b2a0-fd817741e739', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd2d44211-6856-42ed-8f9f-3b1a58b31349', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 961.780796] env[62965]: DEBUG oslo.service.loopingcall [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 961.783386] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 961.783804] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cd6da443-a5fa-484f-9c36-32c7213801df {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.807597] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 961.807597] env[62965]: value = "task-1390153" [ 961.807597] env[62965]: _type = "Task" [ 961.807597] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.815958] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390153, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.891459] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a263c28-47f4-42ad-ab2b-58eb07b6c903 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.900227] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b862b3bb-309f-447e-8e8b-e98bfa24eb9a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.931749] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e22f00aa-6002-4c16-b7e4-b73e451b8a84 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.939649] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf122f1b-9ce3-4eb4-a7de-56c62ddfffed {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.953402] env[62965]: DEBUG nova.compute.provider_tree [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 961.962750] env[62965]: DEBUG oslo_vmware.api [None req-7a739b05-b8d3-437c-bbd4-3851e7b35db0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390148, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.995146] env[62965]: DEBUG oslo_vmware.api [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390152, 'name': Rename_Task, 'duration_secs': 0.14893} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.995390] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 961.995913] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f933c6f8-f5f2-430a-8ff4-711fe9e93a27 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.003337] env[62965]: DEBUG oslo_vmware.api [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 962.003337] env[62965]: value = "task-1390154" [ 962.003337] env[62965]: _type = "Task" [ 962.003337] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.012846] env[62965]: DEBUG oslo_vmware.api [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390154, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.318333] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390153, 'name': CreateVM_Task} progress is 25%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.458403] env[62965]: DEBUG nova.network.neutron [req-c143305f-b926-49f9-835c-f8f9c80da69c req-4a780641-e2cf-4cba-b5e3-ca1faf90a80b service nova] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Updated VIF entry in instance network info cache for port d2d44211-6856-42ed-8f9f-3b1a58b31349. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 962.458826] env[62965]: DEBUG nova.network.neutron [req-c143305f-b926-49f9-835c-f8f9c80da69c req-4a780641-e2cf-4cba-b5e3-ca1faf90a80b service nova] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Updating instance_info_cache with network_info: [{"id": "d2d44211-6856-42ed-8f9f-3b1a58b31349", "address": "fa:16:3e:3b:f0:e9", "network": {"id": "49dd42b0-6397-4fdd-a319-ed920c72a5af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-373172566-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.230", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d91c99b14018469e8998fd1ccc0fab4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2d44211-68", "ovs_interfaceid": "d2d44211-6856-42ed-8f9f-3b1a58b31349", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.469997] env[62965]: DEBUG oslo_vmware.api [None req-7a739b05-b8d3-437c-bbd4-3851e7b35db0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390148, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.488424] env[62965]: DEBUG nova.scheduler.client.report [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Updated inventory for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with generation 107 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 962.488637] env[62965]: DEBUG nova.compute.provider_tree [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Updating resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 generation from 107 to 108 during operation: update_inventory {{(pid=62965) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 962.488788] env[62965]: DEBUG nova.compute.provider_tree [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 962.514115] env[62965]: DEBUG oslo_vmware.api [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390154, 'name': PowerOnVM_Task, 'duration_secs': 0.446876} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.514379] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 962.514583] env[62965]: INFO nova.compute.manager [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Took 6.90 seconds to spawn the instance on the hypervisor. [ 962.514764] env[62965]: DEBUG nova.compute.manager [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 962.515530] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6faaf5ce-725d-4d1c-847f-8b14027758d6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.555466] env[62965]: DEBUG nova.compute.manager [req-326aa618-d528-42ae-bab8-6f0196bb1d59 req-56fd0a14-f0cc-48ae-a1f1-345d2591442d service nova] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Received event network-changed-219e79ce-5ab0-49eb-a38f-6b1ba814cdf3 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 962.555698] env[62965]: DEBUG nova.compute.manager [req-326aa618-d528-42ae-bab8-6f0196bb1d59 req-56fd0a14-f0cc-48ae-a1f1-345d2591442d service nova] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Refreshing instance network info cache due to event network-changed-219e79ce-5ab0-49eb-a38f-6b1ba814cdf3. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 962.555935] env[62965]: DEBUG oslo_concurrency.lockutils [req-326aa618-d528-42ae-bab8-6f0196bb1d59 req-56fd0a14-f0cc-48ae-a1f1-345d2591442d service nova] Acquiring lock "refresh_cache-a63b301d-8d52-4fec-951f-a15cda9449ac" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.556067] env[62965]: DEBUG oslo_concurrency.lockutils [req-326aa618-d528-42ae-bab8-6f0196bb1d59 req-56fd0a14-f0cc-48ae-a1f1-345d2591442d service nova] Acquired lock "refresh_cache-a63b301d-8d52-4fec-951f-a15cda9449ac" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.556254] env[62965]: DEBUG nova.network.neutron [req-326aa618-d528-42ae-bab8-6f0196bb1d59 req-56fd0a14-f0cc-48ae-a1f1-345d2591442d service nova] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Refreshing network info cache for port 219e79ce-5ab0-49eb-a38f-6b1ba814cdf3 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 962.818690] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390153, 'name': CreateVM_Task, 'duration_secs': 0.700297} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.818878] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 962.819526] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c93dd56b-480e-46d6-9aed-4d5aa28dd1e0" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.819697] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c93dd56b-480e-46d6-9aed-4d5aa28dd1e0" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.820085] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/c93dd56b-480e-46d6-9aed-4d5aa28dd1e0" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 962.820403] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3428d55-6c9f-4973-b9e2-81a94ac2ef49 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.824678] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 962.824678] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]528f3ccf-c29c-bd4d-b549-a9dbf6f6137e" [ 962.824678] env[62965]: _type = "Task" [ 962.824678] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.832010] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]528f3ccf-c29c-bd4d-b549-a9dbf6f6137e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.965645] env[62965]: DEBUG oslo_concurrency.lockutils [req-c143305f-b926-49f9-835c-f8f9c80da69c req-4a780641-e2cf-4cba-b5e3-ca1faf90a80b service nova] Releasing lock "refresh_cache-b59d6023-a047-4b32-bd18-1c7cc05f67d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.966172] env[62965]: DEBUG oslo_vmware.api [None req-7a739b05-b8d3-437c-bbd4-3851e7b35db0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390148, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.994178] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62965) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 962.994388] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 4.270s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.994663] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9ec520c8-ed90-4e25-a162-773f48db0ed6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.552s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.994846] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9ec520c8-ed90-4e25-a162-773f48db0ed6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.996853] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.848s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.998423] env[62965]: INFO nova.compute.claims [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 963.020408] env[62965]: INFO nova.scheduler.client.report [None req-9ec520c8-ed90-4e25-a162-773f48db0ed6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Deleted allocations for instance 3223b1ed-abc1-439c-bcad-6317eb204e45 [ 963.032436] env[62965]: INFO nova.compute.manager [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Took 14.18 seconds to build instance. [ 963.253255] env[62965]: DEBUG nova.network.neutron [req-326aa618-d528-42ae-bab8-6f0196bb1d59 req-56fd0a14-f0cc-48ae-a1f1-345d2591442d service nova] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Updated VIF entry in instance network info cache for port 219e79ce-5ab0-49eb-a38f-6b1ba814cdf3. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 963.253629] env[62965]: DEBUG nova.network.neutron [req-326aa618-d528-42ae-bab8-6f0196bb1d59 req-56fd0a14-f0cc-48ae-a1f1-345d2591442d service nova] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Updating instance_info_cache with network_info: [{"id": "219e79ce-5ab0-49eb-a38f-6b1ba814cdf3", "address": "fa:16:3e:6d:19:17", "network": {"id": "98b821d6-efa4-46c8-81e6-a703bd34cb6a", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1649479803-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9a539017d4b84f39a71b6f4098395fbe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8c979f78-8597-41f8-b1de-995014032689", "external-id": "nsx-vlan-transportzone-477", "segmentation_id": 477, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap219e79ce-5a", "ovs_interfaceid": "219e79ce-5ab0-49eb-a38f-6b1ba814cdf3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.335460] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c93dd56b-480e-46d6-9aed-4d5aa28dd1e0" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.335716] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Processing image c93dd56b-480e-46d6-9aed-4d5aa28dd1e0 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 963.335987] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/c93dd56b-480e-46d6-9aed-4d5aa28dd1e0/c93dd56b-480e-46d6-9aed-4d5aa28dd1e0.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.336120] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquired lock "[datastore1] devstack-image-cache_base/c93dd56b-480e-46d6-9aed-4d5aa28dd1e0/c93dd56b-480e-46d6-9aed-4d5aa28dd1e0.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.336302] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 963.336552] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-aae87850-1dab-4bd1-95e7-a5c9a84a6d04 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.357295] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 963.357482] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 963.358316] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4798ede3-4a2f-42a2-941d-138f6708bf09 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.363974] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 963.363974] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52f5db09-75cb-b756-1bd4-994dbc13fc05" [ 963.363974] env[62965]: _type = "Task" [ 963.363974] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.371991] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52f5db09-75cb-b756-1bd4-994dbc13fc05, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.464548] env[62965]: DEBUG oslo_vmware.api [None req-7a739b05-b8d3-437c-bbd4-3851e7b35db0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390148, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.535300] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9ec520c8-ed90-4e25-a162-773f48db0ed6 tempest-ImagesOneServerNegativeTestJSON-872005629 tempest-ImagesOneServerNegativeTestJSON-872005629-project-member] Lock "3223b1ed-abc1-439c-bcad-6317eb204e45" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.842s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.536404] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2e7d2ae1-9e70-4232-a06f-306c0c0a8392 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "08f7b05f-3fb5-48ac-a68a-1d6fd804a622" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.687s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.756743] env[62965]: DEBUG oslo_concurrency.lockutils [req-326aa618-d528-42ae-bab8-6f0196bb1d59 req-56fd0a14-f0cc-48ae-a1f1-345d2591442d service nova] Releasing lock "refresh_cache-a63b301d-8d52-4fec-951f-a15cda9449ac" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.875278] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Preparing fetch location {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 963.875729] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Fetch image to [datastore1] OSTACK_IMG_501b2c94-e452-493d-b7b8-d46eac06195e/OSTACK_IMG_501b2c94-e452-493d-b7b8-d46eac06195e.vmdk {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 963.875729] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Downloading stream optimized image c93dd56b-480e-46d6-9aed-4d5aa28dd1e0 to [datastore1] OSTACK_IMG_501b2c94-e452-493d-b7b8-d46eac06195e/OSTACK_IMG_501b2c94-e452-493d-b7b8-d46eac06195e.vmdk on the data store datastore1 as vApp {{(pid=62965) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 963.875838] env[62965]: DEBUG nova.virt.vmwareapi.images [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Downloading image file data c93dd56b-480e-46d6-9aed-4d5aa28dd1e0 to the ESX as VM named 'OSTACK_IMG_501b2c94-e452-493d-b7b8-d46eac06195e' {{(pid=62965) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 963.967566] env[62965]: DEBUG oslo_vmware.api [None req-7a739b05-b8d3-437c-bbd4-3851e7b35db0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390148, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.968135] env[62965]: DEBUG oslo_vmware.rw_handles [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 963.968135] env[62965]: value = "resgroup-9" [ 963.968135] env[62965]: _type = "ResourcePool" [ 963.968135] env[62965]: }. {{(pid=62965) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 963.968737] env[62965]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-3ec93ca9-03ec-4928-9d0b-28d927fb3431 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.991710] env[62965]: DEBUG oslo_vmware.rw_handles [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lease: (returnval){ [ 963.991710] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]528740a5-aa53-eae4-dacd-429ebba9aea3" [ 963.991710] env[62965]: _type = "HttpNfcLease" [ 963.991710] env[62965]: } obtained for vApp import into resource pool (val){ [ 963.991710] env[62965]: value = "resgroup-9" [ 963.991710] env[62965]: _type = "ResourcePool" [ 963.991710] env[62965]: }. {{(pid=62965) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 963.992133] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the lease: (returnval){ [ 963.992133] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]528740a5-aa53-eae4-dacd-429ebba9aea3" [ 963.992133] env[62965]: _type = "HttpNfcLease" [ 963.992133] env[62965]: } to be ready. {{(pid=62965) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 963.998926] env[62965]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 963.998926] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]528740a5-aa53-eae4-dacd-429ebba9aea3" [ 963.998926] env[62965]: _type = "HttpNfcLease" [ 963.998926] env[62965]: } is initializing. {{(pid=62965) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 964.189639] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8436d84a-4ae1-4682-881e-e151141ebb40 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.197844] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce11ef3e-b080-4772-aa8a-d09a4aa7313c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.227839] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c8d47a1-b3ce-4187-ab5d-443f035c13c1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.242699] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d31f42cd-38f8-4be2-a4a9-9910992d2066 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.265528] env[62965]: DEBUG nova.compute.provider_tree [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 964.469876] env[62965]: DEBUG oslo_vmware.api [None req-7a739b05-b8d3-437c-bbd4-3851e7b35db0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390148, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.501369] env[62965]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 964.501369] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]528740a5-aa53-eae4-dacd-429ebba9aea3" [ 964.501369] env[62965]: _type = "HttpNfcLease" [ 964.501369] env[62965]: } is ready. {{(pid=62965) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 964.501721] env[62965]: DEBUG oslo_vmware.rw_handles [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 964.501721] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]528740a5-aa53-eae4-dacd-429ebba9aea3" [ 964.501721] env[62965]: _type = "HttpNfcLease" [ 964.501721] env[62965]: }. {{(pid=62965) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 964.502712] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de222c6c-4c3f-4b81-810f-e102b39e400c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.515341] env[62965]: DEBUG oslo_vmware.rw_handles [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529f1c0c-932e-f494-8c4e-6ec75adbe6bb/disk-0.vmdk from lease info. {{(pid=62965) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 964.515656] env[62965]: DEBUG oslo_vmware.rw_handles [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Creating HTTP connection to write to file with size = 31666176 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529f1c0c-932e-f494-8c4e-6ec75adbe6bb/disk-0.vmdk. {{(pid=62965) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 964.580888] env[62965]: DEBUG oslo_concurrency.lockutils [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "8a54b436-3b43-439e-bf6d-417db1fa66b2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.581159] env[62965]: DEBUG oslo_concurrency.lockutils [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "8a54b436-3b43-439e-bf6d-417db1fa66b2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.587344] env[62965]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-551c02a0-a824-4d4f-9eca-20237c386dc7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.761106] env[62965]: DEBUG nova.scheduler.client.report [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 964.971603] env[62965]: DEBUG oslo_vmware.api [None req-7a739b05-b8d3-437c-bbd4-3851e7b35db0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390148, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.084959] env[62965]: DEBUG nova.compute.manager [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 965.268174] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.270s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.268782] env[62965]: DEBUG nova.compute.manager [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 965.279445] env[62965]: DEBUG oslo_concurrency.lockutils [None req-17cb7e4a-8c99-4681-89b1-eaa6f3d59e3b tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.366s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.279769] env[62965]: DEBUG oslo_concurrency.lockutils [None req-17cb7e4a-8c99-4681-89b1-eaa6f3d59e3b tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.282709] env[62965]: DEBUG oslo_concurrency.lockutils [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.929s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.283614] env[62965]: INFO nova.compute.claims [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 965.334066] env[62965]: INFO nova.scheduler.client.report [None req-17cb7e4a-8c99-4681-89b1-eaa6f3d59e3b tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Deleted allocations for instance fbbcbe60-92f3-497d-b190-520a2547b3a9 [ 965.475550] env[62965]: DEBUG oslo_vmware.api [None req-7a739b05-b8d3-437c-bbd4-3851e7b35db0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390148, 'name': ReconfigVM_Task, 'duration_secs': 5.835442} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.477995] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7a739b05-b8d3-437c-bbd4-3851e7b35db0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Releasing lock "00f24976-2f0e-4a2f-8f2e-012b0fce479f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.477995] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7a739b05-b8d3-437c-bbd4-3851e7b35db0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Reconfigured VM to detach interface {{(pid=62965) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 965.614661] env[62965]: DEBUG oslo_concurrency.lockutils [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.706473] env[62965]: DEBUG oslo_vmware.rw_handles [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Completed reading data from the image iterator. {{(pid=62965) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 965.706725] env[62965]: DEBUG oslo_vmware.rw_handles [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529f1c0c-932e-f494-8c4e-6ec75adbe6bb/disk-0.vmdk. {{(pid=62965) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 965.707896] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b72f049-99f0-4920-8459-5dd2614bf35f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.716384] env[62965]: DEBUG oslo_vmware.rw_handles [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529f1c0c-932e-f494-8c4e-6ec75adbe6bb/disk-0.vmdk is in state: ready. {{(pid=62965) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 965.716538] env[62965]: DEBUG oslo_vmware.rw_handles [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529f1c0c-932e-f494-8c4e-6ec75adbe6bb/disk-0.vmdk. {{(pid=62965) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 965.716792] env[62965]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-519afc9d-5b05-42b3-8b17-d4edf4df8397 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.778401] env[62965]: DEBUG nova.compute.utils [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 965.780543] env[62965]: DEBUG nova.compute.manager [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 965.780766] env[62965]: DEBUG nova.network.neutron [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 965.843851] env[62965]: DEBUG oslo_concurrency.lockutils [None req-17cb7e4a-8c99-4681-89b1-eaa6f3d59e3b tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "fbbcbe60-92f3-497d-b190-520a2547b3a9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.374s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.850677] env[62965]: DEBUG nova.policy [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '34e15ad3c87445e5a217300a23223540', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8bfa5cfdc7d5496482d87675f2a122dd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 966.131080] env[62965]: DEBUG nova.network.neutron [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Successfully created port: 170ea80e-f3c0-4647-af9d-34ffed538c33 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 966.284207] env[62965]: DEBUG nova.compute.manager [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 966.461553] env[62965]: DEBUG oslo_vmware.rw_handles [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529f1c0c-932e-f494-8c4e-6ec75adbe6bb/disk-0.vmdk. {{(pid=62965) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 966.461800] env[62965]: INFO nova.virt.vmwareapi.images [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Downloaded image file data c93dd56b-480e-46d6-9aed-4d5aa28dd1e0 [ 966.462679] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3be745c-078d-4b52-adab-810c252624aa {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.481979] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2a8bb6d5-4dd7-4d96-ace5-1f7bec2396b2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.507916] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fa6670e-451c-4c6b-a5f5-5271443183fb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.513756] env[62965]: INFO nova.virt.vmwareapi.images [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] The imported VM was unregistered [ 966.516261] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Caching image {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 966.516507] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Creating directory with path [datastore1] devstack-image-cache_base/c93dd56b-480e-46d6-9aed-4d5aa28dd1e0 {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 966.519092] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6b9725e1-5980-4e7d-bc90-fc2624bd341d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.522387] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-879938cf-26ba-4ea8-ad7e-b695f1b41c69 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.563021] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bc87a6c-ec72-4856-81b8-10171a10c750 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.565088] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Created directory with path [datastore1] devstack-image-cache_base/c93dd56b-480e-46d6-9aed-4d5aa28dd1e0 {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 966.565389] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_501b2c94-e452-493d-b7b8-d46eac06195e/OSTACK_IMG_501b2c94-e452-493d-b7b8-d46eac06195e.vmdk to [datastore1] devstack-image-cache_base/c93dd56b-480e-46d6-9aed-4d5aa28dd1e0/c93dd56b-480e-46d6-9aed-4d5aa28dd1e0.vmdk. {{(pid=62965) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 966.566035] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-4dc95347-b1ad-480f-9905-fa3946c7db97 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.577762] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41e5a3ed-3978-42da-b740-af6b853c3645 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.582160] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 966.582160] env[62965]: value = "task-1390157" [ 966.582160] env[62965]: _type = "Task" [ 966.582160] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.594650] env[62965]: DEBUG nova.compute.provider_tree [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 966.602184] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390157, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.941980] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7a739b05-b8d3-437c-bbd4-3851e7b35db0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "refresh_cache-00f24976-2f0e-4a2f-8f2e-012b0fce479f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.941980] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7a739b05-b8d3-437c-bbd4-3851e7b35db0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquired lock "refresh_cache-00f24976-2f0e-4a2f-8f2e-012b0fce479f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.941980] env[62965]: DEBUG nova.network.neutron [None req-7a739b05-b8d3-437c-bbd4-3851e7b35db0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 967.093204] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390157, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.098314] env[62965]: DEBUG nova.scheduler.client.report [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 967.293084] env[62965]: DEBUG nova.compute.manager [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 967.319918] env[62965]: DEBUG nova.virt.hardware [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 967.320276] env[62965]: DEBUG nova.virt.hardware [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 967.320599] env[62965]: DEBUG nova.virt.hardware [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 967.320690] env[62965]: DEBUG nova.virt.hardware [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 967.320777] env[62965]: DEBUG nova.virt.hardware [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 967.320925] env[62965]: DEBUG nova.virt.hardware [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 967.321145] env[62965]: DEBUG nova.virt.hardware [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 967.321333] env[62965]: DEBUG nova.virt.hardware [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 967.321521] env[62965]: DEBUG nova.virt.hardware [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 967.321692] env[62965]: DEBUG nova.virt.hardware [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 967.321896] env[62965]: DEBUG nova.virt.hardware [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 967.322835] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6f8221a-8827-48b0-9dc1-27df81a431eb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.334960] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-504d1842-dcfb-45fc-b38f-21407396ca23 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.440308] env[62965]: DEBUG oslo_concurrency.lockutils [None req-30813006-67b4-4fd7-b4c4-5ef9c82641ae tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "00f24976-2f0e-4a2f-8f2e-012b0fce479f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.440371] env[62965]: DEBUG oslo_concurrency.lockutils [None req-30813006-67b4-4fd7-b4c4-5ef9c82641ae tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "00f24976-2f0e-4a2f-8f2e-012b0fce479f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.440624] env[62965]: DEBUG oslo_concurrency.lockutils [None req-30813006-67b4-4fd7-b4c4-5ef9c82641ae tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "00f24976-2f0e-4a2f-8f2e-012b0fce479f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.440835] env[62965]: DEBUG oslo_concurrency.lockutils [None req-30813006-67b4-4fd7-b4c4-5ef9c82641ae tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "00f24976-2f0e-4a2f-8f2e-012b0fce479f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.441034] env[62965]: DEBUG oslo_concurrency.lockutils [None req-30813006-67b4-4fd7-b4c4-5ef9c82641ae tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "00f24976-2f0e-4a2f-8f2e-012b0fce479f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.445624] env[62965]: INFO nova.compute.manager [None req-30813006-67b4-4fd7-b4c4-5ef9c82641ae tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Terminating instance [ 967.596596] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390157, 'name': MoveVirtualDisk_Task} progress is 18%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.603736] env[62965]: DEBUG oslo_concurrency.lockutils [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.322s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.604304] env[62965]: DEBUG nova.compute.manager [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 967.607339] env[62965]: DEBUG oslo_concurrency.lockutils [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.011s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.607574] env[62965]: DEBUG nova.objects.instance [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lazy-loading 'resources' on Instance uuid a0e9cc1a-bc95-4a44-99c5-aabf85a373f4 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 967.709488] env[62965]: DEBUG nova.compute.manager [req-088a4ce2-8477-4031-8815-b69daccf2fa8 req-b6231fc7-8042-4058-9f09-4b3f9b2b29e7 service nova] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Received event network-vif-plugged-170ea80e-f3c0-4647-af9d-34ffed538c33 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 967.709727] env[62965]: DEBUG oslo_concurrency.lockutils [req-088a4ce2-8477-4031-8815-b69daccf2fa8 req-b6231fc7-8042-4058-9f09-4b3f9b2b29e7 service nova] Acquiring lock "464a3a31-06bf-4662-a5b1-d185d9a64f0b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.710125] env[62965]: DEBUG oslo_concurrency.lockutils [req-088a4ce2-8477-4031-8815-b69daccf2fa8 req-b6231fc7-8042-4058-9f09-4b3f9b2b29e7 service nova] Lock "464a3a31-06bf-4662-a5b1-d185d9a64f0b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.710444] env[62965]: DEBUG oslo_concurrency.lockutils [req-088a4ce2-8477-4031-8815-b69daccf2fa8 req-b6231fc7-8042-4058-9f09-4b3f9b2b29e7 service nova] Lock "464a3a31-06bf-4662-a5b1-d185d9a64f0b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.710680] env[62965]: DEBUG nova.compute.manager [req-088a4ce2-8477-4031-8815-b69daccf2fa8 req-b6231fc7-8042-4058-9f09-4b3f9b2b29e7 service nova] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] No waiting events found dispatching network-vif-plugged-170ea80e-f3c0-4647-af9d-34ffed538c33 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 967.711291] env[62965]: WARNING nova.compute.manager [req-088a4ce2-8477-4031-8815-b69daccf2fa8 req-b6231fc7-8042-4058-9f09-4b3f9b2b29e7 service nova] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Received unexpected event network-vif-plugged-170ea80e-f3c0-4647-af9d-34ffed538c33 for instance with vm_state building and task_state spawning. [ 967.865420] env[62965]: DEBUG nova.network.neutron [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Successfully updated port: 170ea80e-f3c0-4647-af9d-34ffed538c33 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 967.915182] env[62965]: DEBUG nova.compute.manager [req-0dba9a83-1c83-4737-97c6-9ac2de30703d req-3be2a027-c0c5-463b-8f75-fcfe0cc94c9f service nova] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Received event network-changed-170ea80e-f3c0-4647-af9d-34ffed538c33 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 967.915182] env[62965]: DEBUG nova.compute.manager [req-0dba9a83-1c83-4737-97c6-9ac2de30703d req-3be2a027-c0c5-463b-8f75-fcfe0cc94c9f service nova] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Refreshing instance network info cache due to event network-changed-170ea80e-f3c0-4647-af9d-34ffed538c33. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 967.915182] env[62965]: DEBUG oslo_concurrency.lockutils [req-0dba9a83-1c83-4737-97c6-9ac2de30703d req-3be2a027-c0c5-463b-8f75-fcfe0cc94c9f service nova] Acquiring lock "refresh_cache-464a3a31-06bf-4662-a5b1-d185d9a64f0b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.915182] env[62965]: DEBUG oslo_concurrency.lockutils [req-0dba9a83-1c83-4737-97c6-9ac2de30703d req-3be2a027-c0c5-463b-8f75-fcfe0cc94c9f service nova] Acquired lock "refresh_cache-464a3a31-06bf-4662-a5b1-d185d9a64f0b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.915182] env[62965]: DEBUG nova.network.neutron [req-0dba9a83-1c83-4737-97c6-9ac2de30703d req-3be2a027-c0c5-463b-8f75-fcfe0cc94c9f service nova] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Refreshing network info cache for port 170ea80e-f3c0-4647-af9d-34ffed538c33 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 967.950996] env[62965]: DEBUG nova.compute.manager [None req-30813006-67b4-4fd7-b4c4-5ef9c82641ae tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 967.951324] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-30813006-67b4-4fd7-b4c4-5ef9c82641ae tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 967.952421] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-825e35e9-3b63-4ed8-892b-e30b4f848f1a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.966029] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-30813006-67b4-4fd7-b4c4-5ef9c82641ae tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 967.969422] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4eacef28-c4a3-436f-bce0-e81552b29a07 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.983213] env[62965]: DEBUG oslo_vmware.api [None req-30813006-67b4-4fd7-b4c4-5ef9c82641ae tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Waiting for the task: (returnval){ [ 967.983213] env[62965]: value = "task-1390158" [ 967.983213] env[62965]: _type = "Task" [ 967.983213] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.998014] env[62965]: DEBUG oslo_vmware.api [None req-30813006-67b4-4fd7-b4c4-5ef9c82641ae tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390158, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.096058] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390157, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.099958] env[62965]: INFO nova.network.neutron [None req-7a739b05-b8d3-437c-bbd4-3851e7b35db0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Port 9c46a8bf-6b23-45d3-95a4-f0c751895c71 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 968.100435] env[62965]: DEBUG nova.network.neutron [None req-7a739b05-b8d3-437c-bbd4-3851e7b35db0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Updating instance_info_cache with network_info: [{"id": "8fea9d40-d98f-4328-ad3e-ae708a849908", "address": "fa:16:3e:c3:8c:35", "network": {"id": "a9699d8b-5e76-4666-a236-33ab8f22345e", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1329540850-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.162", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1952dedf340f4b7ba0d57af6bba9a749", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a316376e-2ef0-4b1e-b40c-10321ebd7e1a", "external-id": "nsx-vlan-transportzone-942", "segmentation_id": 942, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8fea9d40-d9", "ovs_interfaceid": "8fea9d40-d98f-4328-ad3e-ae708a849908", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.110724] env[62965]: DEBUG nova.compute.utils [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 968.113132] env[62965]: DEBUG nova.compute.manager [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 968.113132] env[62965]: DEBUG nova.network.neutron [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 968.115360] env[62965]: DEBUG nova.objects.instance [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lazy-loading 'numa_topology' on Instance uuid a0e9cc1a-bc95-4a44-99c5-aabf85a373f4 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 968.182354] env[62965]: DEBUG nova.policy [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03c0a9a0ec3c46bf883c68d86cb816d6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '978666785b0743bb843de96585f96fb4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 968.367327] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "refresh_cache-464a3a31-06bf-4662-a5b1-d185d9a64f0b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.461395] env[62965]: DEBUG nova.network.neutron [req-0dba9a83-1c83-4737-97c6-9ac2de30703d req-3be2a027-c0c5-463b-8f75-fcfe0cc94c9f service nova] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 968.497767] env[62965]: DEBUG oslo_vmware.api [None req-30813006-67b4-4fd7-b4c4-5ef9c82641ae tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390158, 'name': PowerOffVM_Task, 'duration_secs': 0.233717} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.499301] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-30813006-67b4-4fd7-b4c4-5ef9c82641ae tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 968.499618] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-30813006-67b4-4fd7-b4c4-5ef9c82641ae tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 968.499878] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a56266dd-066f-4e4a-a2f1-7ae5fafc38e4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.583808] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-30813006-67b4-4fd7-b4c4-5ef9c82641ae tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 968.584114] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-30813006-67b4-4fd7-b4c4-5ef9c82641ae tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Deleting contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 968.584326] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-30813006-67b4-4fd7-b4c4-5ef9c82641ae tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Deleting the datastore file [datastore2] 00f24976-2f0e-4a2f-8f2e-012b0fce479f {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 968.584697] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8824e28b-e61c-4dd9-8efc-d94fc37b163e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.596571] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390157, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.597653] env[62965]: DEBUG nova.network.neutron [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Successfully created port: 2664b4df-ad2e-4114-8837-6c5429ae94e3 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 968.601144] env[62965]: DEBUG oslo_vmware.api [None req-30813006-67b4-4fd7-b4c4-5ef9c82641ae tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Waiting for the task: (returnval){ [ 968.601144] env[62965]: value = "task-1390160" [ 968.601144] env[62965]: _type = "Task" [ 968.601144] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.605754] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7a739b05-b8d3-437c-bbd4-3851e7b35db0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Releasing lock "refresh_cache-00f24976-2f0e-4a2f-8f2e-012b0fce479f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.617555] env[62965]: DEBUG nova.compute.manager [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 968.621652] env[62965]: DEBUG oslo_vmware.api [None req-30813006-67b4-4fd7-b4c4-5ef9c82641ae tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390160, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.622684] env[62965]: DEBUG nova.objects.base [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=62965) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 968.638024] env[62965]: DEBUG nova.network.neutron [req-0dba9a83-1c83-4737-97c6-9ac2de30703d req-3be2a027-c0c5-463b-8f75-fcfe0cc94c9f service nova] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.848371] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c2d2462-9f04-407f-9b81-570ecae8cee4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.859717] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-723e73cd-2651-48b3-922b-ee40d9dffb9d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.902454] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5338f8ae-2b67-4d47-9395-41d26f91ed70 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.913241] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16f12da1-3020-479b-a252-5856069c35c0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.930993] env[62965]: DEBUG nova.compute.provider_tree [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 969.097843] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390157, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.111293] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7a739b05-b8d3-437c-bbd4-3851e7b35db0 tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "interface-00f24976-2f0e-4a2f-8f2e-012b0fce479f-9c46a8bf-6b23-45d3-95a4-f0c751895c71" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.244s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.119972] env[62965]: DEBUG oslo_vmware.api [None req-30813006-67b4-4fd7-b4c4-5ef9c82641ae tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390160, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.138294] env[62965]: DEBUG oslo_concurrency.lockutils [req-0dba9a83-1c83-4737-97c6-9ac2de30703d req-3be2a027-c0c5-463b-8f75-fcfe0cc94c9f service nova] Releasing lock "refresh_cache-464a3a31-06bf-4662-a5b1-d185d9a64f0b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.138647] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquired lock "refresh_cache-464a3a31-06bf-4662-a5b1-d185d9a64f0b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.138647] env[62965]: DEBUG nova.network.neutron [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 969.434705] env[62965]: DEBUG nova.scheduler.client.report [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 969.596159] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390157, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.611376] env[62965]: DEBUG oslo_vmware.api [None req-30813006-67b4-4fd7-b4c4-5ef9c82641ae tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390160, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.53649} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.611796] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-30813006-67b4-4fd7-b4c4-5ef9c82641ae tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 969.611866] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-30813006-67b4-4fd7-b4c4-5ef9c82641ae tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Deleted contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 969.611992] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-30813006-67b4-4fd7-b4c4-5ef9c82641ae tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 969.612181] env[62965]: INFO nova.compute.manager [None req-30813006-67b4-4fd7-b4c4-5ef9c82641ae tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Took 1.66 seconds to destroy the instance on the hypervisor. [ 969.612439] env[62965]: DEBUG oslo.service.loopingcall [None req-30813006-67b4-4fd7-b4c4-5ef9c82641ae tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 969.612674] env[62965]: DEBUG nova.compute.manager [-] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 969.612771] env[62965]: DEBUG nova.network.neutron [-] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 969.636371] env[62965]: DEBUG nova.compute.manager [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 969.664542] env[62965]: DEBUG nova.virt.hardware [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 969.664829] env[62965]: DEBUG nova.virt.hardware [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 969.665019] env[62965]: DEBUG nova.virt.hardware [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 969.665137] env[62965]: DEBUG nova.virt.hardware [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 969.665285] env[62965]: DEBUG nova.virt.hardware [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 969.665446] env[62965]: DEBUG nova.virt.hardware [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 969.665663] env[62965]: DEBUG nova.virt.hardware [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 969.665825] env[62965]: DEBUG nova.virt.hardware [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 969.665997] env[62965]: DEBUG nova.virt.hardware [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 969.666662] env[62965]: DEBUG nova.virt.hardware [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 969.666939] env[62965]: DEBUG nova.virt.hardware [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 969.668059] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12ce6b64-5131-4767-8c7b-f64399542cb8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.681191] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f209f84a-5218-4a24-946d-3634f8f5e748 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.686319] env[62965]: DEBUG nova.network.neutron [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 969.903615] env[62965]: DEBUG nova.network.neutron [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Updating instance_info_cache with network_info: [{"id": "170ea80e-f3c0-4647-af9d-34ffed538c33", "address": "fa:16:3e:23:92:be", "network": {"id": "a5246cc1-6001-4c20-9724-fd36a0437582", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-644702631-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bfa5cfdc7d5496482d87675f2a122dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap170ea80e-f3", "ovs_interfaceid": "170ea80e-f3c0-4647-af9d-34ffed538c33", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.943093] env[62965]: DEBUG oslo_concurrency.lockutils [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.336s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.945891] env[62965]: DEBUG oslo_concurrency.lockutils [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.333s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.946200] env[62965]: DEBUG nova.objects.instance [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lazy-loading 'resources' on Instance uuid 97954de1-7b4b-4615-8d0f-0e8580a515ed {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 970.099419] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390157, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.413561} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.100071] env[62965]: INFO nova.virt.vmwareapi.ds_util [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_501b2c94-e452-493d-b7b8-d46eac06195e/OSTACK_IMG_501b2c94-e452-493d-b7b8-d46eac06195e.vmdk to [datastore1] devstack-image-cache_base/c93dd56b-480e-46d6-9aed-4d5aa28dd1e0/c93dd56b-480e-46d6-9aed-4d5aa28dd1e0.vmdk. [ 970.100071] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Cleaning up location [datastore1] OSTACK_IMG_501b2c94-e452-493d-b7b8-d46eac06195e {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 970.100436] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_501b2c94-e452-493d-b7b8-d46eac06195e {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 970.100760] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d993db3f-7eac-4305-8a59-16323bb82370 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.109155] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 970.109155] env[62965]: value = "task-1390161" [ 970.109155] env[62965]: _type = "Task" [ 970.109155] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.118232] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390161, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.197350] env[62965]: DEBUG nova.compute.manager [req-825ddb3c-5055-48d5-ad8f-06aaf3b35ee3 req-a17feefb-bf4b-44fb-a8c2-f29687c80ab9 service nova] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Received event network-vif-plugged-2664b4df-ad2e-4114-8837-6c5429ae94e3 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 970.197576] env[62965]: DEBUG oslo_concurrency.lockutils [req-825ddb3c-5055-48d5-ad8f-06aaf3b35ee3 req-a17feefb-bf4b-44fb-a8c2-f29687c80ab9 service nova] Acquiring lock "51babf43-b7b0-4731-a989-6977874fb801-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.197798] env[62965]: DEBUG oslo_concurrency.lockutils [req-825ddb3c-5055-48d5-ad8f-06aaf3b35ee3 req-a17feefb-bf4b-44fb-a8c2-f29687c80ab9 service nova] Lock "51babf43-b7b0-4731-a989-6977874fb801-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.197971] env[62965]: DEBUG oslo_concurrency.lockutils [req-825ddb3c-5055-48d5-ad8f-06aaf3b35ee3 req-a17feefb-bf4b-44fb-a8c2-f29687c80ab9 service nova] Lock "51babf43-b7b0-4731-a989-6977874fb801-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.198632] env[62965]: DEBUG nova.compute.manager [req-825ddb3c-5055-48d5-ad8f-06aaf3b35ee3 req-a17feefb-bf4b-44fb-a8c2-f29687c80ab9 service nova] [instance: 51babf43-b7b0-4731-a989-6977874fb801] No waiting events found dispatching network-vif-plugged-2664b4df-ad2e-4114-8837-6c5429ae94e3 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 970.198902] env[62965]: WARNING nova.compute.manager [req-825ddb3c-5055-48d5-ad8f-06aaf3b35ee3 req-a17feefb-bf4b-44fb-a8c2-f29687c80ab9 service nova] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Received unexpected event network-vif-plugged-2664b4df-ad2e-4114-8837-6c5429ae94e3 for instance with vm_state building and task_state spawning. [ 970.364806] env[62965]: DEBUG nova.network.neutron [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Successfully updated port: 2664b4df-ad2e-4114-8837-6c5429ae94e3 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 970.406919] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Releasing lock "refresh_cache-464a3a31-06bf-4662-a5b1-d185d9a64f0b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.407334] env[62965]: DEBUG nova.compute.manager [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Instance network_info: |[{"id": "170ea80e-f3c0-4647-af9d-34ffed538c33", "address": "fa:16:3e:23:92:be", "network": {"id": "a5246cc1-6001-4c20-9724-fd36a0437582", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-644702631-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bfa5cfdc7d5496482d87675f2a122dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap170ea80e-f3", "ovs_interfaceid": "170ea80e-f3c0-4647-af9d-34ffed538c33", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 970.407800] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:23:92:be', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '61b8f0db-488e-42d7-bf6c-6c1665cd5616', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '170ea80e-f3c0-4647-af9d-34ffed538c33', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 970.416496] env[62965]: DEBUG oslo.service.loopingcall [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 970.417414] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 970.417740] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-efaed865-c498-4190-a749-8d228f1337c0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.440815] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 970.440815] env[62965]: value = "task-1390162" [ 970.440815] env[62965]: _type = "Task" [ 970.440815] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.455596] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390162, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.458209] env[62965]: DEBUG oslo_concurrency.lockutils [None req-87baf7c7-17da-46ff-b4f6-392f9dc6654e tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lock "a0e9cc1a-bc95-4a44-99c5-aabf85a373f4" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 35.012s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.459272] env[62965]: DEBUG oslo_concurrency.lockutils [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lock "a0e9cc1a-bc95-4a44-99c5-aabf85a373f4" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 8.785s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.459513] env[62965]: INFO nova.compute.manager [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Unshelving [ 970.625579] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390161, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.12323} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.628491] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 970.628702] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Releasing lock "[datastore1] devstack-image-cache_base/c93dd56b-480e-46d6-9aed-4d5aa28dd1e0/c93dd56b-480e-46d6-9aed-4d5aa28dd1e0.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.628988] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/c93dd56b-480e-46d6-9aed-4d5aa28dd1e0/c93dd56b-480e-46d6-9aed-4d5aa28dd1e0.vmdk to [datastore1] b59d6023-a047-4b32-bd18-1c7cc05f67d7/b59d6023-a047-4b32-bd18-1c7cc05f67d7.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 970.629533] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3da37a74-d95c-4610-be40-cd3f88913d86 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.641024] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 970.641024] env[62965]: value = "task-1390163" [ 970.641024] env[62965]: _type = "Task" [ 970.641024] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.646991] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390163, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.722178] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4e92ad1-60ee-440b-971f-7b8b1ae354b0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.729681] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc987dbd-ba63-4427-9746-0d4244092673 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.764843] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68c27778-f5ec-4ba9-ad3f-c3136e745f6e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.774941] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d477baf3-8f2e-4469-8bb1-d2c5a69b9aa9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.794046] env[62965]: DEBUG nova.compute.provider_tree [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 970.868083] env[62965]: DEBUG oslo_concurrency.lockutils [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "refresh_cache-51babf43-b7b0-4731-a989-6977874fb801" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.868302] env[62965]: DEBUG oslo_concurrency.lockutils [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquired lock "refresh_cache-51babf43-b7b0-4731-a989-6977874fb801" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.869797] env[62965]: DEBUG nova.network.neutron [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 970.955917] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390162, 'name': CreateVM_Task, 'duration_secs': 0.386533} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.956076] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 970.956948] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.957179] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.957682] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 970.958111] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-581d64cf-4558-48d5-8821-d8d1ff131364 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.965831] env[62965]: DEBUG oslo_vmware.api [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 970.965831] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52171e5e-da63-8958-6b8a-06f65ebeb915" [ 970.965831] env[62965]: _type = "Task" [ 970.965831] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.980942] env[62965]: DEBUG oslo_vmware.api [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52171e5e-da63-8958-6b8a-06f65ebeb915, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.017146] env[62965]: DEBUG nova.network.neutron [-] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.152496] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390163, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.299167] env[62965]: DEBUG nova.scheduler.client.report [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 971.411682] env[62965]: DEBUG nova.network.neutron [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 971.488683] env[62965]: DEBUG oslo_vmware.api [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52171e5e-da63-8958-6b8a-06f65ebeb915, 'name': SearchDatastore_Task, 'duration_secs': 0.085264} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.488953] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.489211] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 971.489458] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.489609] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.489794] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 971.491402] env[62965]: DEBUG oslo_concurrency.lockutils [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.491661] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fc5511fa-56b9-4cbc-9e37-339a7d3b8640 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.511881] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 971.512110] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 971.512945] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ef1c88f-d461-4c0d-b683-d92c5f79551f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.519276] env[62965]: INFO nova.compute.manager [-] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Took 1.91 seconds to deallocate network for instance. [ 971.525337] env[62965]: DEBUG oslo_vmware.api [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 971.525337] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5210e899-3fbb-92a3-0392-44b83d0e2907" [ 971.525337] env[62965]: _type = "Task" [ 971.525337] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.537273] env[62965]: DEBUG oslo_vmware.api [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5210e899-3fbb-92a3-0392-44b83d0e2907, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.587920] env[62965]: DEBUG nova.network.neutron [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Updating instance_info_cache with network_info: [{"id": "2664b4df-ad2e-4114-8837-6c5429ae94e3", "address": "fa:16:3e:52:04:a7", "network": {"id": "62052700-1781-4933-93c8-fa1aafdb0fd2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1931798228-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "978666785b0743bb843de96585f96fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2664b4df-ad", "ovs_interfaceid": "2664b4df-ad2e-4114-8837-6c5429ae94e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.652788] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390163, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.809083] env[62965]: DEBUG oslo_concurrency.lockutils [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.863s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.811838] env[62965]: DEBUG oslo_concurrency.lockutils [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.197s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.813490] env[62965]: INFO nova.compute.claims [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 971.833330] env[62965]: INFO nova.scheduler.client.report [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Deleted allocations for instance 97954de1-7b4b-4615-8d0f-0e8580a515ed [ 971.924753] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquiring lock "43db5f53-a719-4384-a8d2-63721dde0187" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.925320] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "43db5f53-a719-4384-a8d2-63721dde0187" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.955699] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquiring lock "fe0559f0-d1ac-4af8-b554-2d31c38c2e73" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.956229] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "fe0559f0-d1ac-4af8-b554-2d31c38c2e73" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.027132] env[62965]: DEBUG oslo_concurrency.lockutils [None req-30813006-67b4-4fd7-b4c4-5ef9c82641ae tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.041070] env[62965]: DEBUG oslo_vmware.api [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5210e899-3fbb-92a3-0392-44b83d0e2907, 'name': SearchDatastore_Task, 'duration_secs': 0.092081} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.041949] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-296a5498-1e73-4793-869e-3b4226e56f3b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.050657] env[62965]: DEBUG oslo_vmware.api [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 972.050657] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52446e76-5a2e-9b9b-d069-061d2ae8b769" [ 972.050657] env[62965]: _type = "Task" [ 972.050657] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.061486] env[62965]: DEBUG oslo_vmware.api [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52446e76-5a2e-9b9b-d069-061d2ae8b769, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.090989] env[62965]: DEBUG oslo_concurrency.lockutils [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Releasing lock "refresh_cache-51babf43-b7b0-4731-a989-6977874fb801" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.091376] env[62965]: DEBUG nova.compute.manager [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Instance network_info: |[{"id": "2664b4df-ad2e-4114-8837-6c5429ae94e3", "address": "fa:16:3e:52:04:a7", "network": {"id": "62052700-1781-4933-93c8-fa1aafdb0fd2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1931798228-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "978666785b0743bb843de96585f96fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2664b4df-ad", "ovs_interfaceid": "2664b4df-ad2e-4114-8837-6c5429ae94e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 972.091814] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:52:04:a7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a10c88d7-d13f-44fd-acee-7a734eb5f56a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2664b4df-ad2e-4114-8837-6c5429ae94e3', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 972.099748] env[62965]: DEBUG oslo.service.loopingcall [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 972.100009] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 972.100316] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-08d026be-78f2-49ad-acf1-6143223fd81a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.124832] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 972.124832] env[62965]: value = "task-1390164" [ 972.124832] env[62965]: _type = "Task" [ 972.124832] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.134655] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390164, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.152962] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390163, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.236568] env[62965]: DEBUG nova.compute.manager [req-04262ea5-eb7c-4166-9e27-0ec52963e1e6 req-4c123cc2-02e7-4632-bc69-f0f6ecdb5f7b service nova] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Received event network-changed-2664b4df-ad2e-4114-8837-6c5429ae94e3 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 972.236697] env[62965]: DEBUG nova.compute.manager [req-04262ea5-eb7c-4166-9e27-0ec52963e1e6 req-4c123cc2-02e7-4632-bc69-f0f6ecdb5f7b service nova] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Refreshing instance network info cache due to event network-changed-2664b4df-ad2e-4114-8837-6c5429ae94e3. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 972.236925] env[62965]: DEBUG oslo_concurrency.lockutils [req-04262ea5-eb7c-4166-9e27-0ec52963e1e6 req-4c123cc2-02e7-4632-bc69-f0f6ecdb5f7b service nova] Acquiring lock "refresh_cache-51babf43-b7b0-4731-a989-6977874fb801" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.237105] env[62965]: DEBUG oslo_concurrency.lockutils [req-04262ea5-eb7c-4166-9e27-0ec52963e1e6 req-4c123cc2-02e7-4632-bc69-f0f6ecdb5f7b service nova] Acquired lock "refresh_cache-51babf43-b7b0-4731-a989-6977874fb801" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.237276] env[62965]: DEBUG nova.network.neutron [req-04262ea5-eb7c-4166-9e27-0ec52963e1e6 req-4c123cc2-02e7-4632-bc69-f0f6ecdb5f7b service nova] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Refreshing network info cache for port 2664b4df-ad2e-4114-8837-6c5429ae94e3 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 972.341390] env[62965]: DEBUG oslo_concurrency.lockutils [None req-08e4dad0-9a56-4065-9314-ea13d5721acd tempest-ImagesTestJSON-375248809 tempest-ImagesTestJSON-375248809-project-member] Lock "97954de1-7b4b-4615-8d0f-0e8580a515ed" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.722s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.429164] env[62965]: DEBUG nova.compute.manager [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 972.462673] env[62965]: DEBUG nova.compute.manager [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 972.564037] env[62965]: DEBUG oslo_vmware.api [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52446e76-5a2e-9b9b-d069-061d2ae8b769, 'name': SearchDatastore_Task, 'duration_secs': 0.086602} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.564354] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.564609] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] 464a3a31-06bf-4662-a5b1-d185d9a64f0b/464a3a31-06bf-4662-a5b1-d185d9a64f0b.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 972.564894] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c9176f20-8a91-42cf-a56b-795bf0be0247 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.577210] env[62965]: DEBUG oslo_vmware.api [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 972.577210] env[62965]: value = "task-1390165" [ 972.577210] env[62965]: _type = "Task" [ 972.577210] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.589370] env[62965]: DEBUG oslo_vmware.api [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390165, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.639796] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390164, 'name': CreateVM_Task} progress is 25%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.653667] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390163, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.971530] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.986160] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.067761] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64ca7758-ecd7-412e-84fd-43be9cbcabcb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.075920] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b056422b-2ae2-4bfa-a1e6-7111b4d4c90d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.088196] env[62965]: DEBUG oslo_vmware.api [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390165, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.117984] env[62965]: DEBUG nova.network.neutron [req-04262ea5-eb7c-4166-9e27-0ec52963e1e6 req-4c123cc2-02e7-4632-bc69-f0f6ecdb5f7b service nova] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Updated VIF entry in instance network info cache for port 2664b4df-ad2e-4114-8837-6c5429ae94e3. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 973.118413] env[62965]: DEBUG nova.network.neutron [req-04262ea5-eb7c-4166-9e27-0ec52963e1e6 req-4c123cc2-02e7-4632-bc69-f0f6ecdb5f7b service nova] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Updating instance_info_cache with network_info: [{"id": "2664b4df-ad2e-4114-8837-6c5429ae94e3", "address": "fa:16:3e:52:04:a7", "network": {"id": "62052700-1781-4933-93c8-fa1aafdb0fd2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1931798228-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "978666785b0743bb843de96585f96fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2664b4df-ad", "ovs_interfaceid": "2664b4df-ad2e-4114-8837-6c5429ae94e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.120109] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7013f978-f1c9-4d93-81f8-9c314294e4cd {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.131573] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70881aca-379a-4ced-93b5-ef201fda9a87 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.147672] env[62965]: DEBUG nova.compute.provider_tree [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 973.154833] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390164, 'name': CreateVM_Task} progress is 99%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.161252] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390163, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.398425} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.161574] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/c93dd56b-480e-46d6-9aed-4d5aa28dd1e0/c93dd56b-480e-46d6-9aed-4d5aa28dd1e0.vmdk to [datastore1] b59d6023-a047-4b32-bd18-1c7cc05f67d7/b59d6023-a047-4b32-bd18-1c7cc05f67d7.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 973.162490] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed0303ae-aff0-4ca3-a583-9705ced0383e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.185368] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] b59d6023-a047-4b32-bd18-1c7cc05f67d7/b59d6023-a047-4b32-bd18-1c7cc05f67d7.vmdk or device None with type streamOptimized {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 973.185947] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d52a5801-c837-4bc3-8965-2e6940733477 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.211610] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 973.211610] env[62965]: value = "task-1390166" [ 973.211610] env[62965]: _type = "Task" [ 973.211610] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.220361] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390166, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.590814] env[62965]: DEBUG oslo_vmware.api [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390165, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.623722] env[62965]: DEBUG oslo_concurrency.lockutils [req-04262ea5-eb7c-4166-9e27-0ec52963e1e6 req-4c123cc2-02e7-4632-bc69-f0f6ecdb5f7b service nova] Releasing lock "refresh_cache-51babf43-b7b0-4731-a989-6977874fb801" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.624064] env[62965]: DEBUG nova.compute.manager [req-04262ea5-eb7c-4166-9e27-0ec52963e1e6 req-4c123cc2-02e7-4632-bc69-f0f6ecdb5f7b service nova] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Received event network-vif-deleted-8fea9d40-d98f-4328-ad3e-ae708a849908 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 973.640930] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390164, 'name': CreateVM_Task} progress is 99%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.689716] env[62965]: ERROR nova.scheduler.client.report [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [req-975ef688-ec05-4018-8b64-4eba05a16ade] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-975ef688-ec05-4018-8b64-4eba05a16ade"}]} [ 973.716913] env[62965]: DEBUG nova.scheduler.client.report [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Refreshing inventories for resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 973.724920] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390166, 'name': ReconfigVM_Task, 'duration_secs': 0.460367} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.725665] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Reconfigured VM instance instance-00000048 to attach disk [datastore1] b59d6023-a047-4b32-bd18-1c7cc05f67d7/b59d6023-a047-4b32-bd18-1c7cc05f67d7.vmdk or device None with type streamOptimized {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 973.727146] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'guest_format': None, 'boot_index': 0, 'encryption_format': None, 'encryption_options': None, 'encrypted': False, 'device_name': '/dev/sda', 'device_type': 'disk', 'size': 0, 'encryption_secret_uuid': None, 'disk_bus': None, 'image_id': '523fe138-8131-44f9-96b3-87f923f60cb8'}], 'ephemerals': [], 'block_device_mapping': [{'delete_on_termination': False, 'guest_format': None, 'boot_index': None, 'mount_device': '/dev/sdb', 'attachment_id': 'd538e9d3-8c0c-435b-89fa-8d324bd35d79', 'device_type': None, 'disk_bus': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-295026', 'volume_id': '437ab7f8-7ad9-4d87-8f07-900702627943', 'name': 'volume-437ab7f8-7ad9-4d87-8f07-900702627943', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'b59d6023-a047-4b32-bd18-1c7cc05f67d7', 'attached_at': '', 'detached_at': '', 'volume_id': '437ab7f8-7ad9-4d87-8f07-900702627943', 'serial': '437ab7f8-7ad9-4d87-8f07-900702627943'}, 'volume_type': None}], 'swap': None} {{(pid=62965) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 973.727591] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Volume attach. Driver type: vmdk {{(pid=62965) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 973.727822] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-295026', 'volume_id': '437ab7f8-7ad9-4d87-8f07-900702627943', 'name': 'volume-437ab7f8-7ad9-4d87-8f07-900702627943', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'b59d6023-a047-4b32-bd18-1c7cc05f67d7', 'attached_at': '', 'detached_at': '', 'volume_id': '437ab7f8-7ad9-4d87-8f07-900702627943', 'serial': '437ab7f8-7ad9-4d87-8f07-900702627943'} {{(pid=62965) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 973.729717] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6742d84d-9aaf-43e9-aedd-b65df9e2086c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.750542] env[62965]: DEBUG nova.scheduler.client.report [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Updating ProviderTree inventory for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 973.750887] env[62965]: DEBUG nova.compute.provider_tree [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 973.753503] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cb4d12d-8cd6-4333-81f2-eff5b94bb7b3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.782648] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] volume-437ab7f8-7ad9-4d87-8f07-900702627943/volume-437ab7f8-7ad9-4d87-8f07-900702627943.vmdk or device None with type thin {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 973.783943] env[62965]: DEBUG nova.scheduler.client.report [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Refreshing aggregate associations for resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8, aggregates: None {{(pid=62965) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 973.787462] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-80fe7a7c-add4-4cb9-b040-e63e62e64474 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.813196] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 973.813196] env[62965]: value = "task-1390167" [ 973.813196] env[62965]: _type = "Task" [ 973.813196] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.822414] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390167, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.825275] env[62965]: DEBUG nova.scheduler.client.report [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Refreshing trait associations for resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64 {{(pid=62965) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 974.060452] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ce6e856-d30d-434e-afd2-24ba7b787798 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.071935] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd7f0a91-a2d0-4e15-9718-19e0d7fd404c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.111329] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c60d761-a590-4b72-b0f3-106dceb50fc7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.117242] env[62965]: DEBUG oslo_vmware.api [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390165, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.163221} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.117868] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] 464a3a31-06bf-4662-a5b1-d185d9a64f0b/464a3a31-06bf-4662-a5b1-d185d9a64f0b.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 974.118120] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 974.118412] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c90d1dd4-4678-4b95-a9f5-2786830ba0e6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.124062] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67187952-f9cc-430b-bacb-1f1ed259cb3e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.129679] env[62965]: DEBUG oslo_vmware.api [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 974.129679] env[62965]: value = "task-1390168" [ 974.129679] env[62965]: _type = "Task" [ 974.129679] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.150511] env[62965]: DEBUG nova.compute.provider_tree [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 974.157503] env[62965]: DEBUG oslo_vmware.api [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390168, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.160410] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390164, 'name': CreateVM_Task} progress is 99%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.323561] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390167, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.640777] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390164, 'name': CreateVM_Task} progress is 99%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.645970] env[62965]: DEBUG oslo_vmware.api [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390168, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.122165} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.646255] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 974.647015] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f991a98-4253-42a0-97b5-1f4dff538958 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.672544] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] 464a3a31-06bf-4662-a5b1-d185d9a64f0b/464a3a31-06bf-4662-a5b1-d185d9a64f0b.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 974.673061] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c52928ef-2b7e-4a94-9d18-667ac77071c0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.695458] env[62965]: DEBUG oslo_vmware.api [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 974.695458] env[62965]: value = "task-1390169" [ 974.695458] env[62965]: _type = "Task" [ 974.695458] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.702677] env[62965]: DEBUG nova.scheduler.client.report [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Updated inventory for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with generation 111 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 974.703103] env[62965]: DEBUG nova.compute.provider_tree [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Updating resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 generation from 111 to 112 during operation: update_inventory {{(pid=62965) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 974.703103] env[62965]: DEBUG nova.compute.provider_tree [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 974.709844] env[62965]: DEBUG oslo_vmware.api [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390169, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.825631] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390167, 'name': ReconfigVM_Task, 'duration_secs': 0.626577} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.826067] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Reconfigured VM instance instance-00000048 to attach disk [datastore1] volume-437ab7f8-7ad9-4d87-8f07-900702627943/volume-437ab7f8-7ad9-4d87-8f07-900702627943.vmdk or device None with type thin {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 974.832382] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f09a7eaf-c42c-4cac-97f1-5d5b2fa250ed {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.852203] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 974.852203] env[62965]: value = "task-1390170" [ 974.852203] env[62965]: _type = "Task" [ 974.852203] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.861213] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390170, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.140893] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390164, 'name': CreateVM_Task} progress is 99%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.206957] env[62965]: DEBUG oslo_vmware.api [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390169, 'name': ReconfigVM_Task, 'duration_secs': 0.462154} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.207744] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Reconfigured VM instance instance-0000005a to attach disk [datastore1] 464a3a31-06bf-4662-a5b1-d185d9a64f0b/464a3a31-06bf-4662-a5b1-d185d9a64f0b.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 975.208215] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dea2ec53-8b41-4f24-9c08-46bf3b60d942 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.212217] env[62965]: DEBUG oslo_concurrency.lockutils [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.400s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.212373] env[62965]: DEBUG nova.compute.manager [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 975.216816] env[62965]: DEBUG oslo_concurrency.lockutils [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.725s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.217058] env[62965]: DEBUG nova.objects.instance [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lazy-loading 'pci_requests' on Instance uuid a0e9cc1a-bc95-4a44-99c5-aabf85a373f4 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 975.218741] env[62965]: DEBUG oslo_vmware.api [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 975.218741] env[62965]: value = "task-1390171" [ 975.218741] env[62965]: _type = "Task" [ 975.218741] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.227553] env[62965]: DEBUG oslo_vmware.api [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390171, 'name': Rename_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.362742] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390170, 'name': ReconfigVM_Task, 'duration_secs': 0.387925} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.363078] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-295026', 'volume_id': '437ab7f8-7ad9-4d87-8f07-900702627943', 'name': 'volume-437ab7f8-7ad9-4d87-8f07-900702627943', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'b59d6023-a047-4b32-bd18-1c7cc05f67d7', 'attached_at': '', 'detached_at': '', 'volume_id': '437ab7f8-7ad9-4d87-8f07-900702627943', 'serial': '437ab7f8-7ad9-4d87-8f07-900702627943'} {{(pid=62965) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 975.363662] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7e222e07-a828-4943-b0db-3fc256b2511d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.370999] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 975.370999] env[62965]: value = "task-1390172" [ 975.370999] env[62965]: _type = "Task" [ 975.370999] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.379175] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390172, 'name': Rename_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.643043] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390164, 'name': CreateVM_Task, 'duration_secs': 3.132487} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.643246] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 975.644033] env[62965]: DEBUG oslo_concurrency.lockutils [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.644209] env[62965]: DEBUG oslo_concurrency.lockutils [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.644547] env[62965]: DEBUG oslo_concurrency.lockutils [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 975.644821] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-160fa469-5c0f-4c1d-8a9f-bd1c24810abb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.650074] env[62965]: DEBUG oslo_vmware.api [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 975.650074] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52a29351-c699-dc8d-2250-fd3780f91403" [ 975.650074] env[62965]: _type = "Task" [ 975.650074] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.658253] env[62965]: DEBUG oslo_vmware.api [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52a29351-c699-dc8d-2250-fd3780f91403, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.721364] env[62965]: DEBUG nova.compute.utils [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 975.724169] env[62965]: DEBUG nova.objects.instance [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lazy-loading 'numa_topology' on Instance uuid a0e9cc1a-bc95-4a44-99c5-aabf85a373f4 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 975.725395] env[62965]: DEBUG nova.compute.manager [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 975.725562] env[62965]: DEBUG nova.network.neutron [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 975.736733] env[62965]: DEBUG oslo_vmware.api [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390171, 'name': Rename_Task, 'duration_secs': 0.186459} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.736980] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 975.737240] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-093006ca-d33f-4304-8cf6-e0a0f857913d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.744854] env[62965]: DEBUG oslo_vmware.api [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 975.744854] env[62965]: value = "task-1390173" [ 975.744854] env[62965]: _type = "Task" [ 975.744854] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.753010] env[62965]: DEBUG oslo_vmware.api [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390173, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.781338] env[62965]: DEBUG nova.policy [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '97ee4fdb7cdb45829c7fe694e883da8c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1ed71548c085499981fa1b102d660368', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 975.881749] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390172, 'name': Rename_Task, 'duration_secs': 0.158952} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.882058] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 975.882303] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8ce1da8d-aecc-410a-8439-5912af9887f6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.888793] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 975.888793] env[62965]: value = "task-1390174" [ 975.888793] env[62965]: _type = "Task" [ 975.888793] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.897507] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390174, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.026463] env[62965]: DEBUG nova.network.neutron [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Successfully created port: a5f75e78-0ab2-4dab-9528-153c25adf153 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 976.173026] env[62965]: DEBUG oslo_vmware.api [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52a29351-c699-dc8d-2250-fd3780f91403, 'name': SearchDatastore_Task, 'duration_secs': 0.021161} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.173026] env[62965]: DEBUG oslo_concurrency.lockutils [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.173026] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 976.173026] env[62965]: DEBUG oslo_concurrency.lockutils [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.173026] env[62965]: DEBUG oslo_concurrency.lockutils [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.173026] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 976.173026] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f7343267-84da-47ac-bf1e-0b68969f958f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.187283] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 976.187283] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 976.187283] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4706519c-2e59-4c13-9622-437c4c874b82 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.195510] env[62965]: DEBUG oslo_vmware.api [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 976.195510] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5259372d-a2a2-c3d8-fe25-e40165b13135" [ 976.195510] env[62965]: _type = "Task" [ 976.195510] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.215044] env[62965]: DEBUG oslo_vmware.api [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5259372d-a2a2-c3d8-fe25-e40165b13135, 'name': SearchDatastore_Task, 'duration_secs': 0.015861} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.216216] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5539905f-a648-4d33-be44-7288f622cd0f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.224435] env[62965]: DEBUG oslo_vmware.api [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 976.224435] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52753cf2-10eb-0d23-ccf0-b9d480bb5ad1" [ 976.224435] env[62965]: _type = "Task" [ 976.224435] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.228455] env[62965]: DEBUG nova.compute.manager [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 976.232494] env[62965]: INFO nova.compute.claims [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 976.243182] env[62965]: DEBUG oslo_vmware.api [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52753cf2-10eb-0d23-ccf0-b9d480bb5ad1, 'name': SearchDatastore_Task, 'duration_secs': 0.011815} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.243565] env[62965]: DEBUG oslo_concurrency.lockutils [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.243897] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] 51babf43-b7b0-4731-a989-6977874fb801/51babf43-b7b0-4731-a989-6977874fb801.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 976.244276] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c5a6f23f-744c-425a-8133-816ed403dc5f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.258888] env[62965]: DEBUG oslo_vmware.api [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390173, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.262021] env[62965]: DEBUG oslo_vmware.api [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 976.262021] env[62965]: value = "task-1390175" [ 976.262021] env[62965]: _type = "Task" [ 976.262021] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.272447] env[62965]: DEBUG oslo_vmware.api [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390175, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.400223] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390174, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.763192] env[62965]: DEBUG oslo_vmware.api [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390173, 'name': PowerOnVM_Task, 'duration_secs': 0.647147} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.766793] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 976.767142] env[62965]: INFO nova.compute.manager [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Took 9.47 seconds to spawn the instance on the hypervisor. [ 976.767297] env[62965]: DEBUG nova.compute.manager [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 976.768076] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-574c410b-871e-4f12-8563-fa3e7dc168bf {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.782622] env[62965]: DEBUG oslo_vmware.api [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390175, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.900060] env[62965]: DEBUG oslo_vmware.api [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390174, 'name': PowerOnVM_Task, 'duration_secs': 0.643552} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.900376] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 977.001178] env[62965]: DEBUG nova.compute.manager [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 977.002267] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a578438-549d-4393-b0bc-c1ddf97af4b9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.243300] env[62965]: DEBUG nova.compute.manager [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 977.268080] env[62965]: DEBUG nova.virt.hardware [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 977.268353] env[62965]: DEBUG nova.virt.hardware [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 977.268510] env[62965]: DEBUG nova.virt.hardware [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 977.268691] env[62965]: DEBUG nova.virt.hardware [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 977.268837] env[62965]: DEBUG nova.virt.hardware [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 977.268983] env[62965]: DEBUG nova.virt.hardware [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 977.269222] env[62965]: DEBUG nova.virt.hardware [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 977.269400] env[62965]: DEBUG nova.virt.hardware [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 977.269570] env[62965]: DEBUG nova.virt.hardware [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 977.269732] env[62965]: DEBUG nova.virt.hardware [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 977.269907] env[62965]: DEBUG nova.virt.hardware [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 977.270707] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-743521f3-e8de-404d-b90b-0d0f16b1dc8b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.280729] env[62965]: DEBUG oslo_vmware.api [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390175, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.710677} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.283131] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] 51babf43-b7b0-4731-a989-6977874fb801/51babf43-b7b0-4731-a989-6977874fb801.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 977.283600] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 977.286948] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-032af2b1-e0af-4bb1-8d40-f8a95b8f79b6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.290013] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-134ec7c5-d4fb-4916-a537-61a30b1a1542 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.297443] env[62965]: INFO nova.compute.manager [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Took 20.16 seconds to build instance. [ 977.314390] env[62965]: DEBUG oslo_vmware.api [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 977.314390] env[62965]: value = "task-1390176" [ 977.314390] env[62965]: _type = "Task" [ 977.314390] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.324154] env[62965]: DEBUG oslo_vmware.api [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390176, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.440578] env[62965]: DEBUG nova.compute.manager [req-7e0c7cd7-91ca-4d2e-bfbb-93f4c67e369d req-4b6d6102-ca8a-4f7e-8e92-422d846b7b15 service nova] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Received event network-vif-plugged-a5f75e78-0ab2-4dab-9528-153c25adf153 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 977.440578] env[62965]: DEBUG oslo_concurrency.lockutils [req-7e0c7cd7-91ca-4d2e-bfbb-93f4c67e369d req-4b6d6102-ca8a-4f7e-8e92-422d846b7b15 service nova] Acquiring lock "8a54b436-3b43-439e-bf6d-417db1fa66b2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.440878] env[62965]: DEBUG oslo_concurrency.lockutils [req-7e0c7cd7-91ca-4d2e-bfbb-93f4c67e369d req-4b6d6102-ca8a-4f7e-8e92-422d846b7b15 service nova] Lock "8a54b436-3b43-439e-bf6d-417db1fa66b2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.440991] env[62965]: DEBUG oslo_concurrency.lockutils [req-7e0c7cd7-91ca-4d2e-bfbb-93f4c67e369d req-4b6d6102-ca8a-4f7e-8e92-422d846b7b15 service nova] Lock "8a54b436-3b43-439e-bf6d-417db1fa66b2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.441231] env[62965]: DEBUG nova.compute.manager [req-7e0c7cd7-91ca-4d2e-bfbb-93f4c67e369d req-4b6d6102-ca8a-4f7e-8e92-422d846b7b15 service nova] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] No waiting events found dispatching network-vif-plugged-a5f75e78-0ab2-4dab-9528-153c25adf153 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 977.441444] env[62965]: WARNING nova.compute.manager [req-7e0c7cd7-91ca-4d2e-bfbb-93f4c67e369d req-4b6d6102-ca8a-4f7e-8e92-422d846b7b15 service nova] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Received unexpected event network-vif-plugged-a5f75e78-0ab2-4dab-9528-153c25adf153 for instance with vm_state building and task_state spawning. [ 977.457797] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5134c53f-784b-429c-9c17-c029d9b9d178 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.466170] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d80c2b5a-3f0c-4c60-82df-07a40cad7b92 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.502068] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cf776e6-84c3-4e90-8502-65a233c45cd3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.508042] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15e2a298-cccf-4021-a381-f50854bdf157 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.518373] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a88acef2-229d-43cd-818b-3c6f4cd363a0 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "b59d6023-a047-4b32-bd18-1c7cc05f67d7" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 32.260s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.527520] env[62965]: DEBUG nova.compute.provider_tree [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 977.551340] env[62965]: DEBUG nova.network.neutron [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Successfully updated port: a5f75e78-0ab2-4dab-9528-153c25adf153 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 977.800489] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c30920de-2935-487b-bb54-5a1f120838b4 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "464a3a31-06bf-4662-a5b1-d185d9a64f0b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.674s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.824997] env[62965]: DEBUG oslo_vmware.api [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390176, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069889} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.825313] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 977.826138] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2a87f9e-a544-4adb-9dfc-4ff1573c97ea {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.848331] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] 51babf43-b7b0-4731-a989-6977874fb801/51babf43-b7b0-4731-a989-6977874fb801.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 977.848601] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a94387ea-e956-49fe-aea1-a42728922450 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.872308] env[62965]: DEBUG oslo_vmware.api [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 977.872308] env[62965]: value = "task-1390177" [ 977.872308] env[62965]: _type = "Task" [ 977.872308] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.880636] env[62965]: DEBUG oslo_vmware.api [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390177, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.030781] env[62965]: DEBUG nova.scheduler.client.report [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 978.053845] env[62965]: DEBUG oslo_concurrency.lockutils [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "refresh_cache-8a54b436-3b43-439e-bf6d-417db1fa66b2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.053979] env[62965]: DEBUG oslo_concurrency.lockutils [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquired lock "refresh_cache-8a54b436-3b43-439e-bf6d-417db1fa66b2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.054037] env[62965]: DEBUG nova.network.neutron [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 978.382830] env[62965]: DEBUG oslo_vmware.api [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390177, 'name': ReconfigVM_Task, 'duration_secs': 0.300308} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.383207] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Reconfigured VM instance instance-0000005b to attach disk [datastore1] 51babf43-b7b0-4731-a989-6977874fb801/51babf43-b7b0-4731-a989-6977874fb801.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 978.383911] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-19d34516-e19b-432d-9832-0afbbca4fa86 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.391077] env[62965]: DEBUG oslo_vmware.api [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 978.391077] env[62965]: value = "task-1390178" [ 978.391077] env[62965]: _type = "Task" [ 978.391077] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.401315] env[62965]: DEBUG oslo_vmware.api [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390178, 'name': Rename_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.552924] env[62965]: DEBUG oslo_concurrency.lockutils [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.319s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.552924] env[62965]: DEBUG oslo_concurrency.lockutils [None req-30813006-67b4-4fd7-b4c4-5ef9c82641ae tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.512s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.552924] env[62965]: DEBUG nova.objects.instance [None req-30813006-67b4-4fd7-b4c4-5ef9c82641ae tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lazy-loading 'resources' on Instance uuid 00f24976-2f0e-4a2f-8f2e-012b0fce479f {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 978.589206] env[62965]: INFO nova.network.neutron [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Updating port 6ffa156f-6a68-424e-b324-660892b9ad22 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 978.594066] env[62965]: DEBUG nova.network.neutron [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 978.826812] env[62965]: DEBUG nova.network.neutron [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Updating instance_info_cache with network_info: [{"id": "a5f75e78-0ab2-4dab-9528-153c25adf153", "address": "fa:16:3e:19:f1:2c", "network": {"id": "a0a5e605-d580-4b7b-b3c0-5c7395bd5a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1679618017-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ed71548c085499981fa1b102d660368", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5f75e78-0a", "ovs_interfaceid": "a5f75e78-0ab2-4dab-9528-153c25adf153", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.901257] env[62965]: DEBUG oslo_vmware.api [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390178, 'name': Rename_Task, 'duration_secs': 0.153475} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.901438] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 978.901689] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cb591390-166a-42ed-b669-0af3b829d173 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.909457] env[62965]: DEBUG oslo_vmware.api [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 978.909457] env[62965]: value = "task-1390179" [ 978.909457] env[62965]: _type = "Task" [ 978.909457] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.917911] env[62965]: DEBUG oslo_vmware.api [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390179, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.209669] env[62965]: DEBUG nova.compute.manager [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Stashing vm_state: active {{(pid=62965) _prep_resize /opt/stack/nova/nova/compute/manager.py:5953}} [ 979.271370] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a392dfc-feaa-435b-9e5b-9b16d125afb7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.281561] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fde656d9-f70b-4817-9137-ddc097925130 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.315942] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74cd1362-829c-4e0e-bf54-f49962781527 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.324383] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fd24dfc-b3c5-45ce-b9ed-747bacdf1497 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.328743] env[62965]: DEBUG oslo_concurrency.lockutils [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Releasing lock "refresh_cache-8a54b436-3b43-439e-bf6d-417db1fa66b2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.329031] env[62965]: DEBUG nova.compute.manager [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Instance network_info: |[{"id": "a5f75e78-0ab2-4dab-9528-153c25adf153", "address": "fa:16:3e:19:f1:2c", "network": {"id": "a0a5e605-d580-4b7b-b3c0-5c7395bd5a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1679618017-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ed71548c085499981fa1b102d660368", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5f75e78-0a", "ovs_interfaceid": "a5f75e78-0ab2-4dab-9528-153c25adf153", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 979.329433] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:19:f1:2c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0cd5d325-3053-407e-a4ee-f627e82a23f9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a5f75e78-0ab2-4dab-9528-153c25adf153', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 979.337129] env[62965]: DEBUG oslo.service.loopingcall [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 979.337959] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 979.338322] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fea9fd45-13e9-4ab7-b239-3d60c2e0b5f4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.369753] env[62965]: DEBUG nova.compute.provider_tree [None req-30813006-67b4-4fd7-b4c4-5ef9c82641ae tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 979.378997] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 979.378997] env[62965]: value = "task-1390180" [ 979.378997] env[62965]: _type = "Task" [ 979.378997] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.392098] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390180, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.420725] env[62965]: DEBUG oslo_vmware.api [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390179, 'name': PowerOnVM_Task, 'duration_secs': 0.486653} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.421018] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 979.421270] env[62965]: INFO nova.compute.manager [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Took 9.78 seconds to spawn the instance on the hypervisor. [ 979.421429] env[62965]: DEBUG nova.compute.manager [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 979.422277] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f240713b-c31d-4080-982a-11a88a25bc1d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.471038] env[62965]: DEBUG nova.compute.manager [req-23838154-9bb2-4bf6-8fca-1f98faec6d8c req-255a2725-fe75-4403-b2e0-a105cccdada0 service nova] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Received event network-changed-a5f75e78-0ab2-4dab-9528-153c25adf153 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 979.471146] env[62965]: DEBUG nova.compute.manager [req-23838154-9bb2-4bf6-8fca-1f98faec6d8c req-255a2725-fe75-4403-b2e0-a105cccdada0 service nova] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Refreshing instance network info cache due to event network-changed-a5f75e78-0ab2-4dab-9528-153c25adf153. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 979.471357] env[62965]: DEBUG oslo_concurrency.lockutils [req-23838154-9bb2-4bf6-8fca-1f98faec6d8c req-255a2725-fe75-4403-b2e0-a105cccdada0 service nova] Acquiring lock "refresh_cache-8a54b436-3b43-439e-bf6d-417db1fa66b2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.471487] env[62965]: DEBUG oslo_concurrency.lockutils [req-23838154-9bb2-4bf6-8fca-1f98faec6d8c req-255a2725-fe75-4403-b2e0-a105cccdada0 service nova] Acquired lock "refresh_cache-8a54b436-3b43-439e-bf6d-417db1fa66b2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.471652] env[62965]: DEBUG nova.network.neutron [req-23838154-9bb2-4bf6-8fca-1f98faec6d8c req-255a2725-fe75-4403-b2e0-a105cccdada0 service nova] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Refreshing network info cache for port a5f75e78-0ab2-4dab-9528-153c25adf153 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 979.729389] env[62965]: DEBUG oslo_concurrency.lockutils [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.873931] env[62965]: DEBUG nova.scheduler.client.report [None req-30813006-67b4-4fd7-b4c4-5ef9c82641ae tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 979.890177] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390180, 'name': CreateVM_Task} progress is 25%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.939193] env[62965]: INFO nova.compute.manager [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Took 20.60 seconds to build instance. [ 980.289871] env[62965]: DEBUG nova.network.neutron [req-23838154-9bb2-4bf6-8fca-1f98faec6d8c req-255a2725-fe75-4403-b2e0-a105cccdada0 service nova] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Updated VIF entry in instance network info cache for port a5f75e78-0ab2-4dab-9528-153c25adf153. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 980.290263] env[62965]: DEBUG nova.network.neutron [req-23838154-9bb2-4bf6-8fca-1f98faec6d8c req-255a2725-fe75-4403-b2e0-a105cccdada0 service nova] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Updating instance_info_cache with network_info: [{"id": "a5f75e78-0ab2-4dab-9528-153c25adf153", "address": "fa:16:3e:19:f1:2c", "network": {"id": "a0a5e605-d580-4b7b-b3c0-5c7395bd5a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1679618017-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ed71548c085499981fa1b102d660368", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5f75e78-0a", "ovs_interfaceid": "a5f75e78-0ab2-4dab-9528-153c25adf153", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.306820] env[62965]: DEBUG oslo_concurrency.lockutils [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquiring lock "refresh_cache-a0e9cc1a-bc95-4a44-99c5-aabf85a373f4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.307012] env[62965]: DEBUG oslo_concurrency.lockutils [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquired lock "refresh_cache-a0e9cc1a-bc95-4a44-99c5-aabf85a373f4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.307199] env[62965]: DEBUG nova.network.neutron [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 980.378642] env[62965]: DEBUG oslo_concurrency.lockutils [None req-30813006-67b4-4fd7-b4c4-5ef9c82641ae tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.840s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.381010] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.410s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.383388] env[62965]: INFO nova.compute.claims [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 980.394921] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390180, 'name': CreateVM_Task, 'duration_secs': 0.845294} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.394921] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 980.395926] env[62965]: DEBUG oslo_concurrency.lockutils [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.395926] env[62965]: DEBUG oslo_concurrency.lockutils [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.396160] env[62965]: DEBUG oslo_concurrency.lockutils [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 980.396374] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5338e00-9880-4049-a801-701eb2331751 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.400956] env[62965]: INFO nova.scheduler.client.report [None req-30813006-67b4-4fd7-b4c4-5ef9c82641ae tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Deleted allocations for instance 00f24976-2f0e-4a2f-8f2e-012b0fce479f [ 980.403123] env[62965]: DEBUG oslo_vmware.api [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 980.403123] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52dbb9c2-7816-9543-a9ee-0ae170b57141" [ 980.403123] env[62965]: _type = "Task" [ 980.403123] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.415722] env[62965]: DEBUG oslo_vmware.api [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52dbb9c2-7816-9543-a9ee-0ae170b57141, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.440878] env[62965]: DEBUG oslo_concurrency.lockutils [None req-28e7694c-3b5f-4a6c-b5d3-c395cf57a8dc tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "51babf43-b7b0-4731-a989-6977874fb801" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.114s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.793171] env[62965]: DEBUG oslo_concurrency.lockutils [req-23838154-9bb2-4bf6-8fca-1f98faec6d8c req-255a2725-fe75-4403-b2e0-a105cccdada0 service nova] Releasing lock "refresh_cache-8a54b436-3b43-439e-bf6d-417db1fa66b2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.916688] env[62965]: DEBUG oslo_vmware.api [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52dbb9c2-7816-9543-a9ee-0ae170b57141, 'name': SearchDatastore_Task, 'duration_secs': 0.014054} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.917149] env[62965]: DEBUG oslo_concurrency.lockutils [None req-30813006-67b4-4fd7-b4c4-5ef9c82641ae tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "00f24976-2f0e-4a2f-8f2e-012b0fce479f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.477s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.918068] env[62965]: DEBUG oslo_concurrency.lockutils [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.918281] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 980.918540] env[62965]: DEBUG oslo_concurrency.lockutils [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.918655] env[62965]: DEBUG oslo_concurrency.lockutils [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.918830] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 980.919251] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-312e3b43-4414-48c4-8390-273dc75cce31 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.927748] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 980.927933] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 980.929418] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-25977daa-9e43-4ced-8da4-853893992402 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.934961] env[62965]: DEBUG oslo_vmware.api [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 980.934961] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52c0d09c-9faa-e9f2-1f1d-cc6247a6efed" [ 980.934961] env[62965]: _type = "Task" [ 980.934961] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.942592] env[62965]: DEBUG oslo_vmware.api [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52c0d09c-9faa-e9f2-1f1d-cc6247a6efed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.018148] env[62965]: DEBUG nova.network.neutron [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Updating instance_info_cache with network_info: [{"id": "6ffa156f-6a68-424e-b324-660892b9ad22", "address": "fa:16:3e:ed:d8:a1", "network": {"id": "e9a6ebdd-0479-45af-b947-5d35ae182c87", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-385052062-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89fd5275b2064288b44e79365f6f0271", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ba07329-1d3e-4ba8-8774-d029262318c4", "external-id": "nsx-vlan-transportzone-534", "segmentation_id": 534, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ffa156f-6a", "ovs_interfaceid": "6ffa156f-6a68-424e-b324-660892b9ad22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.356805] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "51babf43-b7b0-4731-a989-6977874fb801" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.357155] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "51babf43-b7b0-4731-a989-6977874fb801" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.357302] env[62965]: INFO nova.compute.manager [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Shelving [ 981.390531] env[62965]: DEBUG oslo_concurrency.lockutils [None req-73b09f49-9626-4362-88a1-5eb1278fbafb tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "d038608c-9264-4fd5-b9ba-0653c98ec0e4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.390792] env[62965]: DEBUG oslo_concurrency.lockutils [None req-73b09f49-9626-4362-88a1-5eb1278fbafb tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "d038608c-9264-4fd5-b9ba-0653c98ec0e4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.391032] env[62965]: DEBUG oslo_concurrency.lockutils [None req-73b09f49-9626-4362-88a1-5eb1278fbafb tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "d038608c-9264-4fd5-b9ba-0653c98ec0e4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.391615] env[62965]: DEBUG oslo_concurrency.lockutils [None req-73b09f49-9626-4362-88a1-5eb1278fbafb tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "d038608c-9264-4fd5-b9ba-0653c98ec0e4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.391809] env[62965]: DEBUG oslo_concurrency.lockutils [None req-73b09f49-9626-4362-88a1-5eb1278fbafb tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "d038608c-9264-4fd5-b9ba-0653c98ec0e4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.396201] env[62965]: INFO nova.compute.manager [None req-73b09f49-9626-4362-88a1-5eb1278fbafb tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Terminating instance [ 981.453430] env[62965]: DEBUG oslo_vmware.api [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52c0d09c-9faa-e9f2-1f1d-cc6247a6efed, 'name': SearchDatastore_Task, 'duration_secs': 0.00997} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.457884] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ddd50414-4bb3-4a00-9243-5e9a46f8f4e1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.465661] env[62965]: DEBUG oslo_vmware.api [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 981.465661] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5231a8a9-7afc-c8df-6f65-3abbb4faec88" [ 981.465661] env[62965]: _type = "Task" [ 981.465661] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.473550] env[62965]: DEBUG oslo_vmware.api [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5231a8a9-7afc-c8df-6f65-3abbb4faec88, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.495894] env[62965]: DEBUG nova.compute.manager [req-94b0437b-9662-4aff-9222-c196eaae164a req-ef8cf2a7-875c-4719-8461-9a051b7244c3 service nova] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Received event network-vif-plugged-6ffa156f-6a68-424e-b324-660892b9ad22 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 981.496158] env[62965]: DEBUG oslo_concurrency.lockutils [req-94b0437b-9662-4aff-9222-c196eaae164a req-ef8cf2a7-875c-4719-8461-9a051b7244c3 service nova] Acquiring lock "a0e9cc1a-bc95-4a44-99c5-aabf85a373f4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.496339] env[62965]: DEBUG oslo_concurrency.lockutils [req-94b0437b-9662-4aff-9222-c196eaae164a req-ef8cf2a7-875c-4719-8461-9a051b7244c3 service nova] Lock "a0e9cc1a-bc95-4a44-99c5-aabf85a373f4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.496505] env[62965]: DEBUG oslo_concurrency.lockutils [req-94b0437b-9662-4aff-9222-c196eaae164a req-ef8cf2a7-875c-4719-8461-9a051b7244c3 service nova] Lock "a0e9cc1a-bc95-4a44-99c5-aabf85a373f4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.496670] env[62965]: DEBUG nova.compute.manager [req-94b0437b-9662-4aff-9222-c196eaae164a req-ef8cf2a7-875c-4719-8461-9a051b7244c3 service nova] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] No waiting events found dispatching network-vif-plugged-6ffa156f-6a68-424e-b324-660892b9ad22 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 981.496833] env[62965]: WARNING nova.compute.manager [req-94b0437b-9662-4aff-9222-c196eaae164a req-ef8cf2a7-875c-4719-8461-9a051b7244c3 service nova] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Received unexpected event network-vif-plugged-6ffa156f-6a68-424e-b324-660892b9ad22 for instance with vm_state shelved_offloaded and task_state spawning. [ 981.496991] env[62965]: DEBUG nova.compute.manager [req-94b0437b-9662-4aff-9222-c196eaae164a req-ef8cf2a7-875c-4719-8461-9a051b7244c3 service nova] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Received event network-changed-6ffa156f-6a68-424e-b324-660892b9ad22 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 981.497156] env[62965]: DEBUG nova.compute.manager [req-94b0437b-9662-4aff-9222-c196eaae164a req-ef8cf2a7-875c-4719-8461-9a051b7244c3 service nova] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Refreshing instance network info cache due to event network-changed-6ffa156f-6a68-424e-b324-660892b9ad22. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 981.497332] env[62965]: DEBUG oslo_concurrency.lockutils [req-94b0437b-9662-4aff-9222-c196eaae164a req-ef8cf2a7-875c-4719-8461-9a051b7244c3 service nova] Acquiring lock "refresh_cache-a0e9cc1a-bc95-4a44-99c5-aabf85a373f4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.521495] env[62965]: DEBUG oslo_concurrency.lockutils [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Releasing lock "refresh_cache-a0e9cc1a-bc95-4a44-99c5-aabf85a373f4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.523734] env[62965]: DEBUG oslo_concurrency.lockutils [req-94b0437b-9662-4aff-9222-c196eaae164a req-ef8cf2a7-875c-4719-8461-9a051b7244c3 service nova] Acquired lock "refresh_cache-a0e9cc1a-bc95-4a44-99c5-aabf85a373f4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.523734] env[62965]: DEBUG nova.network.neutron [req-94b0437b-9662-4aff-9222-c196eaae164a req-ef8cf2a7-875c-4719-8461-9a051b7244c3 service nova] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Refreshing network info cache for port 6ffa156f-6a68-424e-b324-660892b9ad22 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 981.548760] env[62965]: DEBUG nova.virt.hardware [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='22d3db7c28e6289fd8c6c81a656fbf91',container_format='bare',created_at=2024-10-31T12:37:42Z,direct_url=,disk_format='vmdk',id=c1230705-5bdc-4c15-a740-13db3313def2,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-90516810-shelved',owner='89fd5275b2064288b44e79365f6f0271',properties=ImageMetaProps,protected=,size=31667712,status='active',tags=,updated_at=2024-10-31T12:37:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 981.549006] env[62965]: DEBUG nova.virt.hardware [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 981.549173] env[62965]: DEBUG nova.virt.hardware [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 981.549357] env[62965]: DEBUG nova.virt.hardware [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 981.549505] env[62965]: DEBUG nova.virt.hardware [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 981.549651] env[62965]: DEBUG nova.virt.hardware [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 981.549855] env[62965]: DEBUG nova.virt.hardware [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 981.550017] env[62965]: DEBUG nova.virt.hardware [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 981.550188] env[62965]: DEBUG nova.virt.hardware [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 981.550411] env[62965]: DEBUG nova.virt.hardware [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 981.550596] env[62965]: DEBUG nova.virt.hardware [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 981.551800] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5b94766-fe2f-4ecd-bde6-55cf523d210a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.559505] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-128ca667-9237-4474-9f5c-170d867fcb3e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.575073] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ed:d8:a1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5ba07329-1d3e-4ba8-8774-d029262318c4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6ffa156f-6a68-424e-b324-660892b9ad22', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 981.583530] env[62965]: DEBUG oslo.service.loopingcall [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 981.585913] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 981.586324] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-036e2f1e-ca65-4433-9a95-dcd662aadcbe {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.608642] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 981.608642] env[62965]: value = "task-1390181" [ 981.608642] env[62965]: _type = "Task" [ 981.608642] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.611840] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37f10ba0-bff1-4e00-ba55-0d84743df987 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.618496] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390181, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.620926] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d3faaf7-ebc2-4e54-a4ee-2633e45b7cb8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.650931] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf35085d-e85b-4854-b246-242b3c338108 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.657822] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bd3c752-07c9-4b8d-98e7-c0b57010322a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.671726] env[62965]: DEBUG nova.compute.provider_tree [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 981.901102] env[62965]: DEBUG nova.compute.manager [None req-73b09f49-9626-4362-88a1-5eb1278fbafb tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 981.901284] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-73b09f49-9626-4362-88a1-5eb1278fbafb tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 981.902204] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c03d9610-06a8-4533-a456-38c6e9949a24 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.910746] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-73b09f49-9626-4362-88a1-5eb1278fbafb tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 981.910984] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1636cf34-f6af-46fb-8359-c5ee3306b76e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.916767] env[62965]: DEBUG oslo_vmware.api [None req-73b09f49-9626-4362-88a1-5eb1278fbafb tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Waiting for the task: (returnval){ [ 981.916767] env[62965]: value = "task-1390182" [ 981.916767] env[62965]: _type = "Task" [ 981.916767] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.924644] env[62965]: DEBUG oslo_vmware.api [None req-73b09f49-9626-4362-88a1-5eb1278fbafb tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390182, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.975233] env[62965]: DEBUG oslo_vmware.api [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5231a8a9-7afc-c8df-6f65-3abbb4faec88, 'name': SearchDatastore_Task, 'duration_secs': 0.010811} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.975508] env[62965]: DEBUG oslo_concurrency.lockutils [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.975764] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 8a54b436-3b43-439e-bf6d-417db1fa66b2/8a54b436-3b43-439e-bf6d-417db1fa66b2.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 981.976038] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4b37c61e-9237-42fd-8837-5443d8e6ad80 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.982224] env[62965]: DEBUG oslo_vmware.api [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 981.982224] env[62965]: value = "task-1390183" [ 981.982224] env[62965]: _type = "Task" [ 981.982224] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.991076] env[62965]: DEBUG oslo_vmware.api [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390183, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.118810] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390181, 'name': CreateVM_Task, 'duration_secs': 0.328589} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.119082] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 982.119889] env[62965]: DEBUG oslo_concurrency.lockutils [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1230705-5bdc-4c15-a740-13db3313def2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.120131] env[62965]: DEBUG oslo_concurrency.lockutils [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1230705-5bdc-4c15-a740-13db3313def2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.120588] env[62965]: DEBUG oslo_concurrency.lockutils [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c1230705-5bdc-4c15-a740-13db3313def2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 982.123208] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4acc8b03-c79a-43ae-af0f-cd761988dc7e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.125384] env[62965]: DEBUG oslo_vmware.api [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 982.125384] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5223b70f-294d-ba13-9dd3-cad22291bd97" [ 982.125384] env[62965]: _type = "Task" [ 982.125384] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.133849] env[62965]: DEBUG oslo_vmware.api [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5223b70f-294d-ba13-9dd3-cad22291bd97, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.175236] env[62965]: DEBUG nova.scheduler.client.report [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 982.259261] env[62965]: DEBUG nova.network.neutron [req-94b0437b-9662-4aff-9222-c196eaae164a req-ef8cf2a7-875c-4719-8461-9a051b7244c3 service nova] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Updated VIF entry in instance network info cache for port 6ffa156f-6a68-424e-b324-660892b9ad22. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 982.259734] env[62965]: DEBUG nova.network.neutron [req-94b0437b-9662-4aff-9222-c196eaae164a req-ef8cf2a7-875c-4719-8461-9a051b7244c3 service nova] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Updating instance_info_cache with network_info: [{"id": "6ffa156f-6a68-424e-b324-660892b9ad22", "address": "fa:16:3e:ed:d8:a1", "network": {"id": "e9a6ebdd-0479-45af-b947-5d35ae182c87", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-385052062-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89fd5275b2064288b44e79365f6f0271", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ba07329-1d3e-4ba8-8774-d029262318c4", "external-id": "nsx-vlan-transportzone-534", "segmentation_id": 534, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ffa156f-6a", "ovs_interfaceid": "6ffa156f-6a68-424e-b324-660892b9ad22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.366511] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 982.366837] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ad1153bd-2ba9-4eef-81d2-ad8f27544947 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.375055] env[62965]: DEBUG oslo_vmware.api [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 982.375055] env[62965]: value = "task-1390184" [ 982.375055] env[62965]: _type = "Task" [ 982.375055] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.386226] env[62965]: DEBUG oslo_vmware.api [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390184, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.428010] env[62965]: DEBUG oslo_vmware.api [None req-73b09f49-9626-4362-88a1-5eb1278fbafb tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390182, 'name': PowerOffVM_Task, 'duration_secs': 0.171767} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.428310] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-73b09f49-9626-4362-88a1-5eb1278fbafb tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 982.428500] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-73b09f49-9626-4362-88a1-5eb1278fbafb tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 982.428798] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4e58d5e5-3f50-4346-8b61-54a19e9e55c0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.494600] env[62965]: DEBUG oslo_vmware.api [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390183, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.536684] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-73b09f49-9626-4362-88a1-5eb1278fbafb tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 982.537023] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-73b09f49-9626-4362-88a1-5eb1278fbafb tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Deleting contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 982.537303] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-73b09f49-9626-4362-88a1-5eb1278fbafb tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Deleting the datastore file [datastore2] d038608c-9264-4fd5-b9ba-0653c98ec0e4 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 982.537679] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2b98c17d-ee51-43f3-b495-275b5db7d790 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.545435] env[62965]: DEBUG oslo_vmware.api [None req-73b09f49-9626-4362-88a1-5eb1278fbafb tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Waiting for the task: (returnval){ [ 982.545435] env[62965]: value = "task-1390186" [ 982.545435] env[62965]: _type = "Task" [ 982.545435] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.555158] env[62965]: DEBUG oslo_vmware.api [None req-73b09f49-9626-4362-88a1-5eb1278fbafb tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390186, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.636024] env[62965]: DEBUG oslo_concurrency.lockutils [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1230705-5bdc-4c15-a740-13db3313def2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.636268] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Processing image c1230705-5bdc-4c15-a740-13db3313def2 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 982.636511] env[62965]: DEBUG oslo_concurrency.lockutils [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c1230705-5bdc-4c15-a740-13db3313def2/c1230705-5bdc-4c15-a740-13db3313def2.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.636660] env[62965]: DEBUG oslo_concurrency.lockutils [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c1230705-5bdc-4c15-a740-13db3313def2/c1230705-5bdc-4c15-a740-13db3313def2.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.636917] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 982.637122] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-09a65fe6-ce15-41fe-be2c-6a886ca126c5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.644810] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 982.645022] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 982.645757] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d73b2f0a-e051-4e51-a643-977f8ec78a2c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.650871] env[62965]: DEBUG oslo_vmware.api [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 982.650871] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]529021b3-6047-f82c-a206-13d226c92cfb" [ 982.650871] env[62965]: _type = "Task" [ 982.650871] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.660230] env[62965]: DEBUG oslo_vmware.api [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]529021b3-6047-f82c-a206-13d226c92cfb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.680350] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.299s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.680948] env[62965]: DEBUG nova.compute.manager [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 982.683545] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.698s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.684902] env[62965]: INFO nova.compute.claims [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 982.761992] env[62965]: DEBUG oslo_concurrency.lockutils [req-94b0437b-9662-4aff-9222-c196eaae164a req-ef8cf2a7-875c-4719-8461-9a051b7244c3 service nova] Releasing lock "refresh_cache-a0e9cc1a-bc95-4a44-99c5-aabf85a373f4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.884798] env[62965]: DEBUG oslo_vmware.api [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390184, 'name': PowerOffVM_Task, 'duration_secs': 0.221112} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.885105] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 982.885940] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45669203-b09a-4b1b-9301-9cb9f24c86bb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.903457] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec0a570a-e6ab-46d4-98f7-5a8941592dbc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.993318] env[62965]: DEBUG oslo_vmware.api [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390183, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.5616} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.993704] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 8a54b436-3b43-439e-bf6d-417db1fa66b2/8a54b436-3b43-439e-bf6d-417db1fa66b2.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 982.993931] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 982.994120] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-927269a2-469f-44ad-ad52-c16ee9db709b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.000505] env[62965]: DEBUG oslo_vmware.api [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 983.000505] env[62965]: value = "task-1390187" [ 983.000505] env[62965]: _type = "Task" [ 983.000505] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.008467] env[62965]: DEBUG oslo_vmware.api [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390187, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.054677] env[62965]: DEBUG oslo_vmware.api [None req-73b09f49-9626-4362-88a1-5eb1278fbafb tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Task: {'id': task-1390186, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.16494} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.054954] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-73b09f49-9626-4362-88a1-5eb1278fbafb tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 983.055151] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-73b09f49-9626-4362-88a1-5eb1278fbafb tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Deleted contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 983.055330] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-73b09f49-9626-4362-88a1-5eb1278fbafb tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 983.055547] env[62965]: INFO nova.compute.manager [None req-73b09f49-9626-4362-88a1-5eb1278fbafb tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Took 1.15 seconds to destroy the instance on the hypervisor. [ 983.055799] env[62965]: DEBUG oslo.service.loopingcall [None req-73b09f49-9626-4362-88a1-5eb1278fbafb tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 983.055996] env[62965]: DEBUG nova.compute.manager [-] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 983.056130] env[62965]: DEBUG nova.network.neutron [-] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 983.162584] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Preparing fetch location {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 983.162897] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Fetch image to [datastore2] OSTACK_IMG_5e8d15cc-5d13-4b00-bbf2-807e0e849871/OSTACK_IMG_5e8d15cc-5d13-4b00-bbf2-807e0e849871.vmdk {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 983.163104] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Downloading stream optimized image c1230705-5bdc-4c15-a740-13db3313def2 to [datastore2] OSTACK_IMG_5e8d15cc-5d13-4b00-bbf2-807e0e849871/OSTACK_IMG_5e8d15cc-5d13-4b00-bbf2-807e0e849871.vmdk on the data store datastore2 as vApp {{(pid=62965) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 983.163285] env[62965]: DEBUG nova.virt.vmwareapi.images [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Downloading image file data c1230705-5bdc-4c15-a740-13db3313def2 to the ESX as VM named 'OSTACK_IMG_5e8d15cc-5d13-4b00-bbf2-807e0e849871' {{(pid=62965) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 983.202556] env[62965]: DEBUG nova.compute.utils [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 983.205812] env[62965]: DEBUG nova.compute.manager [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 983.205812] env[62965]: DEBUG nova.network.neutron [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 983.242329] env[62965]: DEBUG oslo_vmware.rw_handles [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 983.242329] env[62965]: value = "resgroup-9" [ 983.242329] env[62965]: _type = "ResourcePool" [ 983.242329] env[62965]: }. {{(pid=62965) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 983.242646] env[62965]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-811eafdb-eb2c-4e0f-af13-088b6edea9cb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.270548] env[62965]: DEBUG oslo_vmware.rw_handles [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lease: (returnval){ [ 983.270548] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52186593-5297-819a-bfe6-8dc09f1fc0f7" [ 983.270548] env[62965]: _type = "HttpNfcLease" [ 983.270548] env[62965]: } obtained for vApp import into resource pool (val){ [ 983.270548] env[62965]: value = "resgroup-9" [ 983.270548] env[62965]: _type = "ResourcePool" [ 983.270548] env[62965]: }. {{(pid=62965) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 983.270890] env[62965]: DEBUG oslo_vmware.api [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the lease: (returnval){ [ 983.270890] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52186593-5297-819a-bfe6-8dc09f1fc0f7" [ 983.270890] env[62965]: _type = "HttpNfcLease" [ 983.270890] env[62965]: } to be ready. {{(pid=62965) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 983.280481] env[62965]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 983.280481] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52186593-5297-819a-bfe6-8dc09f1fc0f7" [ 983.280481] env[62965]: _type = "HttpNfcLease" [ 983.280481] env[62965]: } is initializing. {{(pid=62965) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 983.413366] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Creating Snapshot of the VM instance {{(pid=62965) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 983.413751] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-c9cffc51-f9f5-45f5-8942-d692fb430acb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.421425] env[62965]: DEBUG oslo_vmware.api [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 983.421425] env[62965]: value = "task-1390189" [ 983.421425] env[62965]: _type = "Task" [ 983.421425] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.429345] env[62965]: DEBUG oslo_vmware.api [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390189, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.459938] env[62965]: DEBUG nova.policy [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e7510e1ec5ee4253b67be988b6b5ab51', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5d2d71b59ede430f846520bf49249e4a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 983.512815] env[62965]: DEBUG oslo_vmware.api [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390187, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.291507} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.514354] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 983.515662] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f611f5b-376d-4dc1-a99b-c716668e1897 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.540423] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] 8a54b436-3b43-439e-bf6d-417db1fa66b2/8a54b436-3b43-439e-bf6d-417db1fa66b2.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 983.540750] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-27286102-ef9b-46bf-bab1-e664505b0742 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.562403] env[62965]: DEBUG oslo_vmware.api [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 983.562403] env[62965]: value = "task-1390190" [ 983.562403] env[62965]: _type = "Task" [ 983.562403] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.570612] env[62965]: DEBUG oslo_vmware.api [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390190, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.607882] env[62965]: DEBUG nova.compute.manager [req-9cf0f04b-1097-4bc7-9ae1-129f8a7e58f4 req-e6dd8985-61f3-4db6-92b7-5cbda4eac9b1 service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Received event network-vif-deleted-b1718061-edf1-4077-b57f-0e54aea365a6 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 983.608108] env[62965]: INFO nova.compute.manager [req-9cf0f04b-1097-4bc7-9ae1-129f8a7e58f4 req-e6dd8985-61f3-4db6-92b7-5cbda4eac9b1 service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Neutron deleted interface b1718061-edf1-4077-b57f-0e54aea365a6; detaching it from the instance and deleting it from the info cache [ 983.608338] env[62965]: DEBUG nova.network.neutron [req-9cf0f04b-1097-4bc7-9ae1-129f8a7e58f4 req-e6dd8985-61f3-4db6-92b7-5cbda4eac9b1 service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.710034] env[62965]: DEBUG nova.compute.manager [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 983.780671] env[62965]: DEBUG nova.network.neutron [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Successfully created port: b70e2be1-f890-4164-9e52-1e4cceaa0be5 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 983.787016] env[62965]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 983.787016] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52186593-5297-819a-bfe6-8dc09f1fc0f7" [ 983.787016] env[62965]: _type = "HttpNfcLease" [ 983.787016] env[62965]: } is initializing. {{(pid=62965) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 983.929333] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79749c2f-05a9-431a-9d8e-bc2a2ecabe27 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.934928] env[62965]: DEBUG oslo_vmware.api [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390189, 'name': CreateSnapshot_Task, 'duration_secs': 0.487897} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.935584] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Created Snapshot of the VM instance {{(pid=62965) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 983.936311] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-796c7a23-0d7f-4427-9824-890be57e4f11 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.941703] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddf34af2-e16a-4f7a-b971-46d64360f733 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.986024] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b63b327-31a3-4f88-ad11-8354eafd1c0c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.991979] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceb859b7-507a-4ecc-bf67-795f4e417dd1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.006386] env[62965]: DEBUG nova.compute.provider_tree [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 984.071707] env[62965]: DEBUG oslo_vmware.api [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390190, 'name': ReconfigVM_Task, 'duration_secs': 0.34452} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.071994] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Reconfigured VM instance instance-0000005c to attach disk [datastore2] 8a54b436-3b43-439e-bf6d-417db1fa66b2/8a54b436-3b43-439e-bf6d-417db1fa66b2.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 984.072651] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b9f3805c-ebb0-48d7-9b69-893ae6482663 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.077790] env[62965]: DEBUG nova.network.neutron [-] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 984.080168] env[62965]: DEBUG oslo_vmware.api [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 984.080168] env[62965]: value = "task-1390191" [ 984.080168] env[62965]: _type = "Task" [ 984.080168] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.089350] env[62965]: DEBUG oslo_vmware.api [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390191, 'name': Rename_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.112302] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4bf49e61-06ee-40b5-89aa-0565bc1580eb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.123369] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd5233a9-49f0-4a30-a1f4-c9555bba2504 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.157051] env[62965]: DEBUG nova.compute.manager [req-9cf0f04b-1097-4bc7-9ae1-129f8a7e58f4 req-e6dd8985-61f3-4db6-92b7-5cbda4eac9b1 service nova] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Detach interface failed, port_id=b1718061-edf1-4077-b57f-0e54aea365a6, reason: Instance d038608c-9264-4fd5-b9ba-0653c98ec0e4 could not be found. {{(pid=62965) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11294}} [ 984.279260] env[62965]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 984.279260] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52186593-5297-819a-bfe6-8dc09f1fc0f7" [ 984.279260] env[62965]: _type = "HttpNfcLease" [ 984.279260] env[62965]: } is ready. {{(pid=62965) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 984.279598] env[62965]: DEBUG oslo_vmware.rw_handles [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 984.279598] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52186593-5297-819a-bfe6-8dc09f1fc0f7" [ 984.279598] env[62965]: _type = "HttpNfcLease" [ 984.279598] env[62965]: }. {{(pid=62965) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 984.280410] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3d3d0f8-fb4b-4cf2-adcf-7b4bd1a4a13d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.287816] env[62965]: DEBUG oslo_vmware.rw_handles [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ca5247-4e72-720d-7d7c-c88c72112cf5/disk-0.vmdk from lease info. {{(pid=62965) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 984.287999] env[62965]: DEBUG oslo_vmware.rw_handles [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Creating HTTP connection to write to file with size = 31667712 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ca5247-4e72-720d-7d7c-c88c72112cf5/disk-0.vmdk. {{(pid=62965) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 984.353439] env[62965]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-1d9fc92b-0142-4eb5-82ab-ee1795798484 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.458434] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Creating linked-clone VM from snapshot {{(pid=62965) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 984.460052] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-40f29b72-a68d-4625-9432-e21ecad6e8bc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.471670] env[62965]: DEBUG oslo_vmware.api [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 984.471670] env[62965]: value = "task-1390192" [ 984.471670] env[62965]: _type = "Task" [ 984.471670] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.481980] env[62965]: DEBUG oslo_vmware.api [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390192, 'name': CloneVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.526967] env[62965]: ERROR nova.scheduler.client.report [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [req-b0ff0194-55ce-4c3d-85d0-916c4b85a06f] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-b0ff0194-55ce-4c3d-85d0-916c4b85a06f"}]} [ 984.582061] env[62965]: INFO nova.compute.manager [-] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Took 1.53 seconds to deallocate network for instance. [ 984.598589] env[62965]: DEBUG oslo_vmware.api [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390191, 'name': Rename_Task, 'duration_secs': 0.151269} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.598932] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 984.599291] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c5b73f0b-749e-4ebd-ad17-c42e56163f69 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.607027] env[62965]: DEBUG oslo_vmware.api [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 984.607027] env[62965]: value = "task-1390193" [ 984.607027] env[62965]: _type = "Task" [ 984.607027] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.619523] env[62965]: DEBUG oslo_vmware.api [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390193, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.634543] env[62965]: DEBUG nova.scheduler.client.report [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Refreshing inventories for resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 984.652389] env[62965]: DEBUG nova.scheduler.client.report [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Updating ProviderTree inventory for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 984.652654] env[62965]: DEBUG nova.compute.provider_tree [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 984.666750] env[62965]: DEBUG nova.scheduler.client.report [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Refreshing aggregate associations for resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8, aggregates: None {{(pid=62965) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 984.686271] env[62965]: DEBUG nova.scheduler.client.report [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Refreshing trait associations for resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64 {{(pid=62965) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 984.723355] env[62965]: DEBUG nova.compute.manager [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 984.754046] env[62965]: DEBUG nova.virt.hardware [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 984.754315] env[62965]: DEBUG nova.virt.hardware [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 984.754472] env[62965]: DEBUG nova.virt.hardware [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 984.754670] env[62965]: DEBUG nova.virt.hardware [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 984.754820] env[62965]: DEBUG nova.virt.hardware [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 984.754971] env[62965]: DEBUG nova.virt.hardware [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 984.755197] env[62965]: DEBUG nova.virt.hardware [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 984.755359] env[62965]: DEBUG nova.virt.hardware [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 984.755531] env[62965]: DEBUG nova.virt.hardware [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 984.755695] env[62965]: DEBUG nova.virt.hardware [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 984.755865] env[62965]: DEBUG nova.virt.hardware [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 984.760375] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-908f7014-4916-4c0c-9026-574b88a19f22 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.775692] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f88a0356-e105-41e4-add1-5517ec93f5bb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.920857] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05df7359-8989-468d-810c-ba157065d4ae {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.930975] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b497b9bf-0ff9-414c-b83e-fde2fe1aaf0a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.969412] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-956d9db0-a984-4c6f-aefa-760047e099c5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.983648] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aec8187-b0cf-4d7f-931c-b16c1000e969 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.987787] env[62965]: DEBUG oslo_vmware.api [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390192, 'name': CloneVM_Task} progress is 94%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.002979] env[62965]: DEBUG nova.compute.provider_tree [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 985.095272] env[62965]: DEBUG oslo_concurrency.lockutils [None req-73b09f49-9626-4362-88a1-5eb1278fbafb tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.118998] env[62965]: DEBUG oslo_vmware.api [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390193, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.350472] env[62965]: DEBUG nova.network.neutron [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Successfully updated port: b70e2be1-f890-4164-9e52-1e4cceaa0be5 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 985.484234] env[62965]: DEBUG oslo_vmware.api [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390192, 'name': CloneVM_Task} progress is 94%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.544111] env[62965]: DEBUG nova.scheduler.client.report [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Updated inventory for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with generation 114 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 985.544472] env[62965]: DEBUG nova.compute.provider_tree [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Updating resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 generation from 114 to 115 during operation: update_inventory {{(pid=62965) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 985.544719] env[62965]: DEBUG nova.compute.provider_tree [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 985.616873] env[62965]: DEBUG oslo_vmware.api [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390193, 'name': PowerOnVM_Task, 'duration_secs': 0.730394} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.617171] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 985.617382] env[62965]: INFO nova.compute.manager [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Took 8.37 seconds to spawn the instance on the hypervisor. [ 985.617559] env[62965]: DEBUG nova.compute.manager [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 985.618363] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7440eae7-8870-4523-920d-a27dfe006c5e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.634623] env[62965]: DEBUG nova.compute.manager [req-99920a6b-9fab-4223-92b7-5ca7c7312dfd req-27f73eba-58b9-4567-9700-6f07096a89b6 service nova] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Received event network-vif-plugged-b70e2be1-f890-4164-9e52-1e4cceaa0be5 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 985.634740] env[62965]: DEBUG oslo_concurrency.lockutils [req-99920a6b-9fab-4223-92b7-5ca7c7312dfd req-27f73eba-58b9-4567-9700-6f07096a89b6 service nova] Acquiring lock "43db5f53-a719-4384-a8d2-63721dde0187-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.634963] env[62965]: DEBUG oslo_concurrency.lockutils [req-99920a6b-9fab-4223-92b7-5ca7c7312dfd req-27f73eba-58b9-4567-9700-6f07096a89b6 service nova] Lock "43db5f53-a719-4384-a8d2-63721dde0187-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.635040] env[62965]: DEBUG oslo_concurrency.lockutils [req-99920a6b-9fab-4223-92b7-5ca7c7312dfd req-27f73eba-58b9-4567-9700-6f07096a89b6 service nova] Lock "43db5f53-a719-4384-a8d2-63721dde0187-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.635215] env[62965]: DEBUG nova.compute.manager [req-99920a6b-9fab-4223-92b7-5ca7c7312dfd req-27f73eba-58b9-4567-9700-6f07096a89b6 service nova] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] No waiting events found dispatching network-vif-plugged-b70e2be1-f890-4164-9e52-1e4cceaa0be5 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 985.635364] env[62965]: WARNING nova.compute.manager [req-99920a6b-9fab-4223-92b7-5ca7c7312dfd req-27f73eba-58b9-4567-9700-6f07096a89b6 service nova] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Received unexpected event network-vif-plugged-b70e2be1-f890-4164-9e52-1e4cceaa0be5 for instance with vm_state building and task_state spawning. [ 985.635527] env[62965]: DEBUG nova.compute.manager [req-99920a6b-9fab-4223-92b7-5ca7c7312dfd req-27f73eba-58b9-4567-9700-6f07096a89b6 service nova] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Received event network-changed-b70e2be1-f890-4164-9e52-1e4cceaa0be5 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 985.635684] env[62965]: DEBUG nova.compute.manager [req-99920a6b-9fab-4223-92b7-5ca7c7312dfd req-27f73eba-58b9-4567-9700-6f07096a89b6 service nova] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Refreshing instance network info cache due to event network-changed-b70e2be1-f890-4164-9e52-1e4cceaa0be5. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 985.635871] env[62965]: DEBUG oslo_concurrency.lockutils [req-99920a6b-9fab-4223-92b7-5ca7c7312dfd req-27f73eba-58b9-4567-9700-6f07096a89b6 service nova] Acquiring lock "refresh_cache-43db5f53-a719-4384-a8d2-63721dde0187" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.636019] env[62965]: DEBUG oslo_concurrency.lockutils [req-99920a6b-9fab-4223-92b7-5ca7c7312dfd req-27f73eba-58b9-4567-9700-6f07096a89b6 service nova] Acquired lock "refresh_cache-43db5f53-a719-4384-a8d2-63721dde0187" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.636179] env[62965]: DEBUG nova.network.neutron [req-99920a6b-9fab-4223-92b7-5ca7c7312dfd req-27f73eba-58b9-4567-9700-6f07096a89b6 service nova] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Refreshing network info cache for port b70e2be1-f890-4164-9e52-1e4cceaa0be5 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 985.746299] env[62965]: DEBUG oslo_vmware.rw_handles [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Completed reading data from the image iterator. {{(pid=62965) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 985.746564] env[62965]: DEBUG oslo_vmware.rw_handles [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ca5247-4e72-720d-7d7c-c88c72112cf5/disk-0.vmdk. {{(pid=62965) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 985.747575] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3ced09a-6112-4e06-8155-254f8a587148 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.754289] env[62965]: DEBUG oslo_vmware.rw_handles [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ca5247-4e72-720d-7d7c-c88c72112cf5/disk-0.vmdk is in state: ready. {{(pid=62965) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 985.754436] env[62965]: DEBUG oslo_vmware.rw_handles [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ca5247-4e72-720d-7d7c-c88c72112cf5/disk-0.vmdk. {{(pid=62965) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 985.754630] env[62965]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-d929a463-8be8-49a9-8ad1-7357ad0039fd {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.853706] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquiring lock "refresh_cache-43db5f53-a719-4384-a8d2-63721dde0187" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.983320] env[62965]: DEBUG oslo_vmware.api [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390192, 'name': CloneVM_Task, 'duration_secs': 1.275502} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.984140] env[62965]: INFO nova.virt.vmwareapi.vmops [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Created linked-clone VM from snapshot [ 985.984471] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3b6bd0c-a739-4cc1-adda-c839316cf638 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.992471] env[62965]: DEBUG nova.virt.vmwareapi.images [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Uploading image a792f1cf-fa3a-46b4-9178-12b6f8866058 {{(pid=62965) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 986.017113] env[62965]: DEBUG oslo_vmware.rw_handles [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 986.017113] env[62965]: value = "vm-295039" [ 986.017113] env[62965]: _type = "VirtualMachine" [ 986.017113] env[62965]: }. {{(pid=62965) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 986.017433] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-02e2d4a1-6ec4-4907-89c4-c6fb8b378e47 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.024494] env[62965]: DEBUG oslo_vmware.rw_handles [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lease: (returnval){ [ 986.024494] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52bcec72-51e4-5d02-8126-b05f209e5717" [ 986.024494] env[62965]: _type = "HttpNfcLease" [ 986.024494] env[62965]: } obtained for exporting VM: (result){ [ 986.024494] env[62965]: value = "vm-295039" [ 986.024494] env[62965]: _type = "VirtualMachine" [ 986.024494] env[62965]: }. {{(pid=62965) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 986.024810] env[62965]: DEBUG oslo_vmware.api [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the lease: (returnval){ [ 986.024810] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52bcec72-51e4-5d02-8126-b05f209e5717" [ 986.024810] env[62965]: _type = "HttpNfcLease" [ 986.024810] env[62965]: } to be ready. {{(pid=62965) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 986.031193] env[62965]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 986.031193] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52bcec72-51e4-5d02-8126-b05f209e5717" [ 986.031193] env[62965]: _type = "HttpNfcLease" [ 986.031193] env[62965]: } is initializing. {{(pid=62965) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 986.050939] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.367s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.051456] env[62965]: DEBUG nova.compute.manager [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 986.054040] env[62965]: DEBUG oslo_concurrency.lockutils [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 6.325s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.135711] env[62965]: INFO nova.compute.manager [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Took 20.55 seconds to build instance. [ 986.166833] env[62965]: DEBUG nova.network.neutron [req-99920a6b-9fab-4223-92b7-5ca7c7312dfd req-27f73eba-58b9-4567-9700-6f07096a89b6 service nova] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 986.242875] env[62965]: DEBUG nova.network.neutron [req-99920a6b-9fab-4223-92b7-5ca7c7312dfd req-27f73eba-58b9-4567-9700-6f07096a89b6 service nova] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.339048] env[62965]: DEBUG oslo_vmware.rw_handles [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52ca5247-4e72-720d-7d7c-c88c72112cf5/disk-0.vmdk. {{(pid=62965) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 986.339171] env[62965]: INFO nova.virt.vmwareapi.images [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Downloaded image file data c1230705-5bdc-4c15-a740-13db3313def2 [ 986.340011] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64edd67e-0fa2-4be7-b407-fc8f8daf118c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.355184] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8f82de7d-8b02-4acc-9ad2-017bcbf71257 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.377828] env[62965]: INFO nova.virt.vmwareapi.images [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] The imported VM was unregistered [ 986.380156] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Caching image {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 986.380447] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Creating directory with path [datastore2] devstack-image-cache_base/c1230705-5bdc-4c15-a740-13db3313def2 {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 986.380720] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-40d8a908-cb5b-4e3d-ac3e-dd16fd0b8ec0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.392576] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Created directory with path [datastore2] devstack-image-cache_base/c1230705-5bdc-4c15-a740-13db3313def2 {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 986.393181] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_5e8d15cc-5d13-4b00-bbf2-807e0e849871/OSTACK_IMG_5e8d15cc-5d13-4b00-bbf2-807e0e849871.vmdk to [datastore2] devstack-image-cache_base/c1230705-5bdc-4c15-a740-13db3313def2/c1230705-5bdc-4c15-a740-13db3313def2.vmdk. {{(pid=62965) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 986.393181] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-f1287260-d496-4f4c-83b1-801979133e7c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.398771] env[62965]: DEBUG oslo_vmware.api [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 986.398771] env[62965]: value = "task-1390196" [ 986.398771] env[62965]: _type = "Task" [ 986.398771] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.405909] env[62965]: DEBUG oslo_vmware.api [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390196, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.533159] env[62965]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 986.533159] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52bcec72-51e4-5d02-8126-b05f209e5717" [ 986.533159] env[62965]: _type = "HttpNfcLease" [ 986.533159] env[62965]: } is ready. {{(pid=62965) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 986.533632] env[62965]: DEBUG oslo_vmware.rw_handles [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 986.533632] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52bcec72-51e4-5d02-8126-b05f209e5717" [ 986.533632] env[62965]: _type = "HttpNfcLease" [ 986.533632] env[62965]: }. {{(pid=62965) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 986.534763] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69008385-ce93-4432-8119-d6673e7578be {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.545143] env[62965]: DEBUG oslo_vmware.rw_handles [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5209d8fd-b8a4-6232-3d48-f3099d427df7/disk-0.vmdk from lease info. {{(pid=62965) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 986.545465] env[62965]: DEBUG oslo_vmware.rw_handles [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5209d8fd-b8a4-6232-3d48-f3099d427df7/disk-0.vmdk for reading. {{(pid=62965) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 986.619107] env[62965]: DEBUG nova.compute.utils [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 986.623219] env[62965]: INFO nova.compute.claims [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 986.628618] env[62965]: DEBUG nova.compute.manager [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 986.628814] env[62965]: DEBUG nova.network.neutron [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 986.637667] env[62965]: DEBUG oslo_concurrency.lockutils [None req-25d94f74-ae8c-4049-b3ff-a077b6551ab2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "8a54b436-3b43-439e-bf6d-417db1fa66b2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.056s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.673805] env[62965]: DEBUG nova.policy [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e7510e1ec5ee4253b67be988b6b5ab51', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5d2d71b59ede430f846520bf49249e4a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 986.686459] env[62965]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-60d0c749-24df-410e-a9ae-3a8cf6da3c41 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.746133] env[62965]: DEBUG oslo_concurrency.lockutils [req-99920a6b-9fab-4223-92b7-5ca7c7312dfd req-27f73eba-58b9-4567-9700-6f07096a89b6 service nova] Releasing lock "refresh_cache-43db5f53-a719-4384-a8d2-63721dde0187" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.746559] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquired lock "refresh_cache-43db5f53-a719-4384-a8d2-63721dde0187" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.746718] env[62965]: DEBUG nova.network.neutron [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 986.911610] env[62965]: DEBUG oslo_vmware.api [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390196, 'name': MoveVirtualDisk_Task} progress is 18%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.960390] env[62965]: DEBUG nova.network.neutron [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Successfully created port: 75d00c8d-3b3f-4a42-b9bb-219bbf284819 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 987.128999] env[62965]: DEBUG nova.compute.manager [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 987.137689] env[62965]: INFO nova.compute.resource_tracker [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Updating resource usage from migration a802f9c1-2aa9-4613-9e17-47bbb76297fb [ 987.284680] env[62965]: DEBUG nova.network.neutron [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 987.386889] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d08de04c-559e-4cf5-a656-7cc7d941bbf3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.399820] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5125b57-00ab-4794-9992-6cc55e07b11c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.417461] env[62965]: DEBUG oslo_vmware.api [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390196, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.445617] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c47460a-5274-4c05-96cc-6a60f3b0018a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.458022] env[62965]: DEBUG nova.network.neutron [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Updating instance_info_cache with network_info: [{"id": "b70e2be1-f890-4164-9e52-1e4cceaa0be5", "address": "fa:16:3e:21:67:69", "network": {"id": "c4c8235c-333c-40e7-8744-110eaa0aa23c", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1256649418-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d2d71b59ede430f846520bf49249e4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb70e2be1-f8", "ovs_interfaceid": "b70e2be1-f890-4164-9e52-1e4cceaa0be5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.460663] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddfd9c11-1ed7-4ac4-aa07-865ae4a537cf {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.480031] env[62965]: DEBUG nova.compute.provider_tree [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 987.483577] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7e1a1167-a069-450d-bb6c-410b0d91a5f1 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "8a54b436-3b43-439e-bf6d-417db1fa66b2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.484174] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7e1a1167-a069-450d-bb6c-410b0d91a5f1 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "8a54b436-3b43-439e-bf6d-417db1fa66b2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.485066] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7e1a1167-a069-450d-bb6c-410b0d91a5f1 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "8a54b436-3b43-439e-bf6d-417db1fa66b2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.485066] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7e1a1167-a069-450d-bb6c-410b0d91a5f1 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "8a54b436-3b43-439e-bf6d-417db1fa66b2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.485066] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7e1a1167-a069-450d-bb6c-410b0d91a5f1 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "8a54b436-3b43-439e-bf6d-417db1fa66b2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.487266] env[62965]: INFO nova.compute.manager [None req-7e1a1167-a069-450d-bb6c-410b0d91a5f1 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Terminating instance [ 987.918147] env[62965]: DEBUG oslo_vmware.api [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390196, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.967098] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Releasing lock "refresh_cache-43db5f53-a719-4384-a8d2-63721dde0187" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.967326] env[62965]: DEBUG nova.compute.manager [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Instance network_info: |[{"id": "b70e2be1-f890-4164-9e52-1e4cceaa0be5", "address": "fa:16:3e:21:67:69", "network": {"id": "c4c8235c-333c-40e7-8744-110eaa0aa23c", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1256649418-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d2d71b59ede430f846520bf49249e4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb70e2be1-f8", "ovs_interfaceid": "b70e2be1-f890-4164-9e52-1e4cceaa0be5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 987.967733] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:21:67:69', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dc6d5964-1106-4345-a26d-185dabd4ff0f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b70e2be1-f890-4164-9e52-1e4cceaa0be5', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 987.976251] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Creating folder: Project (5d2d71b59ede430f846520bf49249e4a). Parent ref: group-v294931. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 987.976612] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-25c757c1-1811-4f23-855e-e2a4f0040d73 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.984150] env[62965]: DEBUG nova.scheduler.client.report [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 987.994174] env[62965]: DEBUG nova.compute.manager [None req-7e1a1167-a069-450d-bb6c-410b0d91a5f1 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 987.994448] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7e1a1167-a069-450d-bb6c-410b0d91a5f1 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 987.994773] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Created folder: Project (5d2d71b59ede430f846520bf49249e4a) in parent group-v294931. [ 987.994997] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Creating folder: Instances. Parent ref: group-v295040. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 987.995894] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dce6adc-1bbf-499e-aae9-91bc37ec5397 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.999063] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fc965bf6-d666-4758-99d9-95f59a2398db {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.011326] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e1a1167-a069-450d-bb6c-410b0d91a5f1 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 988.013081] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-80fd8d20-62a5-4ed7-8747-b3409c4b25fb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.014827] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Created folder: Instances in parent group-v295040. [ 988.015133] env[62965]: DEBUG oslo.service.loopingcall [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 988.015607] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 988.016194] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7fa9c38c-d6bc-4251-96ba-f75a412048eb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.037360] env[62965]: DEBUG oslo_vmware.api [None req-7e1a1167-a069-450d-bb6c-410b0d91a5f1 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 988.037360] env[62965]: value = "task-1390199" [ 988.037360] env[62965]: _type = "Task" [ 988.037360] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.044805] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 988.044805] env[62965]: value = "task-1390200" [ 988.044805] env[62965]: _type = "Task" [ 988.044805] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.048872] env[62965]: DEBUG oslo_vmware.api [None req-7e1a1167-a069-450d-bb6c-410b0d91a5f1 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390199, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.058812] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390200, 'name': CreateVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.140663] env[62965]: DEBUG nova.compute.manager [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 988.169115] env[62965]: DEBUG nova.virt.hardware [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 988.169423] env[62965]: DEBUG nova.virt.hardware [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 988.169523] env[62965]: DEBUG nova.virt.hardware [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 988.169653] env[62965]: DEBUG nova.virt.hardware [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 988.169802] env[62965]: DEBUG nova.virt.hardware [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 988.169951] env[62965]: DEBUG nova.virt.hardware [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 988.170383] env[62965]: DEBUG nova.virt.hardware [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 988.171034] env[62965]: DEBUG nova.virt.hardware [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 988.171034] env[62965]: DEBUG nova.virt.hardware [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 988.171034] env[62965]: DEBUG nova.virt.hardware [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 988.171268] env[62965]: DEBUG nova.virt.hardware [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 988.172012] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94e97de6-6550-4131-bd0d-2587b5d63f1a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.183344] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-987d0402-f1fd-4da2-9e3a-d20256c766f3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.188500] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 988.188770] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 988.425373] env[62965]: DEBUG oslo_vmware.api [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390196, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.434074] env[62965]: DEBUG nova.compute.manager [req-6e561936-eb51-41f6-9f77-4ad435647543 req-1d597cc8-1e6a-4f41-9267-24fe373d3bff service nova] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Received event network-vif-plugged-75d00c8d-3b3f-4a42-b9bb-219bbf284819 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 988.434393] env[62965]: DEBUG oslo_concurrency.lockutils [req-6e561936-eb51-41f6-9f77-4ad435647543 req-1d597cc8-1e6a-4f41-9267-24fe373d3bff service nova] Acquiring lock "fe0559f0-d1ac-4af8-b554-2d31c38c2e73-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.434621] env[62965]: DEBUG oslo_concurrency.lockutils [req-6e561936-eb51-41f6-9f77-4ad435647543 req-1d597cc8-1e6a-4f41-9267-24fe373d3bff service nova] Lock "fe0559f0-d1ac-4af8-b554-2d31c38c2e73-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.434815] env[62965]: DEBUG oslo_concurrency.lockutils [req-6e561936-eb51-41f6-9f77-4ad435647543 req-1d597cc8-1e6a-4f41-9267-24fe373d3bff service nova] Lock "fe0559f0-d1ac-4af8-b554-2d31c38c2e73-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.435045] env[62965]: DEBUG nova.compute.manager [req-6e561936-eb51-41f6-9f77-4ad435647543 req-1d597cc8-1e6a-4f41-9267-24fe373d3bff service nova] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] No waiting events found dispatching network-vif-plugged-75d00c8d-3b3f-4a42-b9bb-219bbf284819 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 988.435285] env[62965]: WARNING nova.compute.manager [req-6e561936-eb51-41f6-9f77-4ad435647543 req-1d597cc8-1e6a-4f41-9267-24fe373d3bff service nova] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Received unexpected event network-vif-plugged-75d00c8d-3b3f-4a42-b9bb-219bbf284819 for instance with vm_state building and task_state spawning. [ 988.489866] env[62965]: DEBUG oslo_concurrency.lockutils [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.436s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.490154] env[62965]: INFO nova.compute.manager [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Migrating [ 988.499102] env[62965]: DEBUG oslo_concurrency.lockutils [None req-73b09f49-9626-4362-88a1-5eb1278fbafb tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.404s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.499419] env[62965]: DEBUG nova.objects.instance [None req-73b09f49-9626-4362-88a1-5eb1278fbafb tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lazy-loading 'resources' on Instance uuid d038608c-9264-4fd5-b9ba-0653c98ec0e4 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 988.552071] env[62965]: DEBUG oslo_vmware.api [None req-7e1a1167-a069-450d-bb6c-410b0d91a5f1 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390199, 'name': PowerOffVM_Task, 'duration_secs': 0.485898} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.556875] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e1a1167-a069-450d-bb6c-410b0d91a5f1 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 988.557190] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7e1a1167-a069-450d-bb6c-410b0d91a5f1 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 988.557607] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-92132118-184f-4a1d-a0b1-74794cd97169 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.566924] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390200, 'name': CreateVM_Task} progress is 25%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.697032] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 988.697032] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Starting heal instance info cache {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10257}} [ 988.702538] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7e1a1167-a069-450d-bb6c-410b0d91a5f1 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 988.702538] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7e1a1167-a069-450d-bb6c-410b0d91a5f1 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Deleting contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 988.702538] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e1a1167-a069-450d-bb6c-410b0d91a5f1 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Deleting the datastore file [datastore2] 8a54b436-3b43-439e-bf6d-417db1fa66b2 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 988.702726] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c8bd05b1-2ded-49e5-8d51-506c01728346 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.711772] env[62965]: DEBUG oslo_vmware.api [None req-7e1a1167-a069-450d-bb6c-410b0d91a5f1 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 988.711772] env[62965]: value = "task-1390202" [ 988.711772] env[62965]: _type = "Task" [ 988.711772] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.722928] env[62965]: DEBUG oslo_vmware.api [None req-7e1a1167-a069-450d-bb6c-410b0d91a5f1 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390202, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.918675] env[62965]: DEBUG oslo_vmware.api [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390196, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.516237} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.918954] env[62965]: INFO nova.virt.vmwareapi.ds_util [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_5e8d15cc-5d13-4b00-bbf2-807e0e849871/OSTACK_IMG_5e8d15cc-5d13-4b00-bbf2-807e0e849871.vmdk to [datastore2] devstack-image-cache_base/c1230705-5bdc-4c15-a740-13db3313def2/c1230705-5bdc-4c15-a740-13db3313def2.vmdk. [ 988.919165] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Cleaning up location [datastore2] OSTACK_IMG_5e8d15cc-5d13-4b00-bbf2-807e0e849871 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 988.919339] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_5e8d15cc-5d13-4b00-bbf2-807e0e849871 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 988.919649] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-04e43d6d-e358-41f6-9dcc-892ced3fd3f9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.926555] env[62965]: DEBUG oslo_vmware.api [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 988.926555] env[62965]: value = "task-1390203" [ 988.926555] env[62965]: _type = "Task" [ 988.926555] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.934647] env[62965]: DEBUG oslo_vmware.api [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390203, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.989645] env[62965]: DEBUG nova.network.neutron [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Successfully updated port: 75d00c8d-3b3f-4a42-b9bb-219bbf284819 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 989.012641] env[62965]: DEBUG oslo_concurrency.lockutils [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "refresh_cache-464a3a31-06bf-4662-a5b1-d185d9a64f0b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.012766] env[62965]: DEBUG oslo_concurrency.lockutils [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquired lock "refresh_cache-464a3a31-06bf-4662-a5b1-d185d9a64f0b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.013104] env[62965]: DEBUG nova.network.neutron [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 989.016102] env[62965]: DEBUG nova.compute.manager [req-d47869bb-1bef-464c-a075-b282a77ce75c req-656af826-38f3-48a7-9c5d-0744edee7794 service nova] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Received event network-changed-75d00c8d-3b3f-4a42-b9bb-219bbf284819 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 989.016294] env[62965]: DEBUG nova.compute.manager [req-d47869bb-1bef-464c-a075-b282a77ce75c req-656af826-38f3-48a7-9c5d-0744edee7794 service nova] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Refreshing instance network info cache due to event network-changed-75d00c8d-3b3f-4a42-b9bb-219bbf284819. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 989.016496] env[62965]: DEBUG oslo_concurrency.lockutils [req-d47869bb-1bef-464c-a075-b282a77ce75c req-656af826-38f3-48a7-9c5d-0744edee7794 service nova] Acquiring lock "refresh_cache-fe0559f0-d1ac-4af8-b554-2d31c38c2e73" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.016717] env[62965]: DEBUG oslo_concurrency.lockutils [req-d47869bb-1bef-464c-a075-b282a77ce75c req-656af826-38f3-48a7-9c5d-0744edee7794 service nova] Acquired lock "refresh_cache-fe0559f0-d1ac-4af8-b554-2d31c38c2e73" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.016972] env[62965]: DEBUG nova.network.neutron [req-d47869bb-1bef-464c-a075-b282a77ce75c req-656af826-38f3-48a7-9c5d-0744edee7794 service nova] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Refreshing network info cache for port 75d00c8d-3b3f-4a42-b9bb-219bbf284819 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 989.060740] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390200, 'name': CreateVM_Task} progress is 99%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.206458] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44965f4c-fd9a-42b8-8e04-86a8a6c8b2aa {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.218078] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df2204c7-69fc-4463-8f3c-31844b947cfa {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.226309] env[62965]: DEBUG oslo_vmware.api [None req-7e1a1167-a069-450d-bb6c-410b0d91a5f1 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390202, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.360868} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.254462] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e1a1167-a069-450d-bb6c-410b0d91a5f1 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 989.254742] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7e1a1167-a069-450d-bb6c-410b0d91a5f1 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Deleted contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 989.255111] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7e1a1167-a069-450d-bb6c-410b0d91a5f1 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 989.255239] env[62965]: INFO nova.compute.manager [None req-7e1a1167-a069-450d-bb6c-410b0d91a5f1 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Took 1.26 seconds to destroy the instance on the hypervisor. [ 989.255513] env[62965]: DEBUG oslo.service.loopingcall [None req-7e1a1167-a069-450d-bb6c-410b0d91a5f1 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 989.256937] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Acquiring lock "refresh_cache-924c5cdc-0450-4ced-b920-50e2d5060fd2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.256937] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Acquired lock "refresh_cache-924c5cdc-0450-4ced-b920-50e2d5060fd2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.257064] env[62965]: DEBUG nova.network.neutron [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Forcefully refreshing network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 989.258346] env[62965]: DEBUG nova.compute.manager [-] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 989.258447] env[62965]: DEBUG nova.network.neutron [-] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 989.261034] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce88860a-a208-4a3a-af86-ebdde2cc6baa {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.270461] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75fd5b46-264d-48d1-a305-1765d30578ce {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.289654] env[62965]: DEBUG nova.compute.provider_tree [None req-73b09f49-9626-4362-88a1-5eb1278fbafb tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 989.437962] env[62965]: DEBUG oslo_vmware.api [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390203, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.080755} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.438567] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 989.438567] env[62965]: DEBUG oslo_concurrency.lockutils [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c1230705-5bdc-4c15-a740-13db3313def2/c1230705-5bdc-4c15-a740-13db3313def2.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.438846] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c1230705-5bdc-4c15-a740-13db3313def2/c1230705-5bdc-4c15-a740-13db3313def2.vmdk to [datastore2] a0e9cc1a-bc95-4a44-99c5-aabf85a373f4/a0e9cc1a-bc95-4a44-99c5-aabf85a373f4.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 989.442021] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-692aa620-5a7c-483a-845b-f73f9a759802 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.447063] env[62965]: DEBUG oslo_vmware.api [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 989.447063] env[62965]: value = "task-1390204" [ 989.447063] env[62965]: _type = "Task" [ 989.447063] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.456373] env[62965]: DEBUG oslo_vmware.api [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390204, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.493566] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquiring lock "refresh_cache-fe0559f0-d1ac-4af8-b554-2d31c38c2e73" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.570501] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390200, 'name': CreateVM_Task, 'duration_secs': 1.061964} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.570685] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 989.571450] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.571588] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.571912] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 989.572196] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-096d7cf4-ea55-4319-b93f-3ea2bdcf335b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.577962] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 989.577962] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52756793-b4fd-d043-0628-12b1909b813f" [ 989.577962] env[62965]: _type = "Task" [ 989.577962] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.588248] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52756793-b4fd-d043-0628-12b1909b813f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.589317] env[62965]: DEBUG nova.network.neutron [req-d47869bb-1bef-464c-a075-b282a77ce75c req-656af826-38f3-48a7-9c5d-0744edee7794 service nova] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 989.765371] env[62965]: DEBUG nova.network.neutron [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Updating instance_info_cache with network_info: [{"id": "170ea80e-f3c0-4647-af9d-34ffed538c33", "address": "fa:16:3e:23:92:be", "network": {"id": "a5246cc1-6001-4c20-9724-fd36a0437582", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-644702631-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bfa5cfdc7d5496482d87675f2a122dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap170ea80e-f3", "ovs_interfaceid": "170ea80e-f3c0-4647-af9d-34ffed538c33", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.774797] env[62965]: DEBUG nova.network.neutron [req-d47869bb-1bef-464c-a075-b282a77ce75c req-656af826-38f3-48a7-9c5d-0744edee7794 service nova] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.835026] env[62965]: DEBUG nova.scheduler.client.report [None req-73b09f49-9626-4362-88a1-5eb1278fbafb tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Updated inventory for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with generation 115 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 989.835026] env[62965]: DEBUG nova.compute.provider_tree [None req-73b09f49-9626-4362-88a1-5eb1278fbafb tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Updating resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 generation from 115 to 116 during operation: update_inventory {{(pid=62965) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 989.835026] env[62965]: DEBUG nova.compute.provider_tree [None req-73b09f49-9626-4362-88a1-5eb1278fbafb tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 989.958836] env[62965]: DEBUG oslo_vmware.api [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390204, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.084749] env[62965]: DEBUG nova.network.neutron [-] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.091249] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52756793-b4fd-d043-0628-12b1909b813f, 'name': SearchDatastore_Task, 'duration_secs': 0.009427} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.091576] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.091990] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 990.092190] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.092265] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.092528] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 990.092781] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-30bb8ab4-c409-40a5-acaa-481635a32f1c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.112060] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 990.112208] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 990.113054] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a4eecfa3-dabb-4006-857d-7bdacf8c0a03 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.123110] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 990.123110] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]528abc31-fe7f-4537-c140-9b469fc28b3b" [ 990.123110] env[62965]: _type = "Task" [ 990.123110] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.135425] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]528abc31-fe7f-4537-c140-9b469fc28b3b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.278068] env[62965]: DEBUG oslo_concurrency.lockutils [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Releasing lock "refresh_cache-464a3a31-06bf-4662-a5b1-d185d9a64f0b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.280121] env[62965]: DEBUG oslo_concurrency.lockutils [req-d47869bb-1bef-464c-a075-b282a77ce75c req-656af826-38f3-48a7-9c5d-0744edee7794 service nova] Releasing lock "refresh_cache-fe0559f0-d1ac-4af8-b554-2d31c38c2e73" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.280507] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquired lock "refresh_cache-fe0559f0-d1ac-4af8-b554-2d31c38c2e73" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.280573] env[62965]: DEBUG nova.network.neutron [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 990.341083] env[62965]: DEBUG oslo_concurrency.lockutils [None req-73b09f49-9626-4362-88a1-5eb1278fbafb tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.842s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.363414] env[62965]: INFO nova.scheduler.client.report [None req-73b09f49-9626-4362-88a1-5eb1278fbafb tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Deleted allocations for instance d038608c-9264-4fd5-b9ba-0653c98ec0e4 [ 990.459312] env[62965]: DEBUG oslo_vmware.api [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390204, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.541611] env[62965]: DEBUG nova.network.neutron [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Updating instance_info_cache with network_info: [{"id": "80afc0d2-0d83-4374-a4c0-792eabf66c40", "address": "fa:16:3e:1c:0d:5e", "network": {"id": "7ef06255-01ec-4620-8f3d-6085875aa8a3", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-118464573-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.223", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c1caebff08644e21b3f1c461aa115b8e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap80afc0d2-0d", "ovs_interfaceid": "80afc0d2-0d83-4374-a4c0-792eabf66c40", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.587327] env[62965]: INFO nova.compute.manager [-] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Took 1.33 seconds to deallocate network for instance. [ 990.636167] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]528abc31-fe7f-4537-c140-9b469fc28b3b, 'name': SearchDatastore_Task, 'duration_secs': 0.084847} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.637117] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6938bb85-0f0f-4cd4-9f47-d72d21a7c301 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.645965] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 990.645965] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52bec89b-25a6-153b-54da-0b2641594a03" [ 990.645965] env[62965]: _type = "Task" [ 990.645965] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.656586] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52bec89b-25a6-153b-54da-0b2641594a03, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.820725] env[62965]: DEBUG nova.network.neutron [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 990.843814] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "8efad776-6cbe-49ab-83dc-a410d8417505" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.844084] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "8efad776-6cbe-49ab-83dc-a410d8417505" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.871234] env[62965]: DEBUG oslo_concurrency.lockutils [None req-73b09f49-9626-4362-88a1-5eb1278fbafb tempest-AttachInterfacesTestJSON-4414725 tempest-AttachInterfacesTestJSON-4414725-project-member] Lock "d038608c-9264-4fd5-b9ba-0653c98ec0e4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.480s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.960612] env[62965]: DEBUG oslo_vmware.api [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390204, 'name': CopyVirtualDisk_Task} progress is 66%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.989435] env[62965]: DEBUG nova.network.neutron [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Updating instance_info_cache with network_info: [{"id": "75d00c8d-3b3f-4a42-b9bb-219bbf284819", "address": "fa:16:3e:1f:ce:31", "network": {"id": "c4c8235c-333c-40e7-8744-110eaa0aa23c", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1256649418-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d2d71b59ede430f846520bf49249e4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75d00c8d-3b", "ovs_interfaceid": "75d00c8d-3b3f-4a42-b9bb-219bbf284819", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.043097] env[62965]: DEBUG nova.compute.manager [req-489f3131-f295-4667-9fdd-65cf6afb146c req-9a6415fa-8669-4a61-b10f-68cbcabe3721 service nova] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Received event network-vif-deleted-a5f75e78-0ab2-4dab-9528-153c25adf153 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 991.044854] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Releasing lock "refresh_cache-924c5cdc-0450-4ced-b920-50e2d5060fd2" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.044854] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Updated the network info_cache for instance {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10328}} [ 991.044983] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 991.045243] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 991.045572] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 991.046114] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 991.046407] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 991.046808] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 991.046970] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62965) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10876}} [ 991.047287] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 991.097058] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7e1a1167-a069-450d-bb6c-410b0d91a5f1 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.097058] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7e1a1167-a069-450d-bb6c-410b0d91a5f1 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.097058] env[62965]: DEBUG nova.objects.instance [None req-7e1a1167-a069-450d-bb6c-410b0d91a5f1 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lazy-loading 'resources' on Instance uuid 8a54b436-3b43-439e-bf6d-417db1fa66b2 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 991.160581] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52bec89b-25a6-153b-54da-0b2641594a03, 'name': SearchDatastore_Task, 'duration_secs': 0.075971} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.160930] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.161334] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 43db5f53-a719-4384-a8d2-63721dde0187/43db5f53-a719-4384-a8d2-63721dde0187.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 991.161658] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8486117a-306a-499b-b419-fe18c4d3c405 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.172677] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 991.172677] env[62965]: value = "task-1390205" [ 991.172677] env[62965]: _type = "Task" [ 991.172677] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.185703] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390205, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.347138] env[62965]: DEBUG nova.compute.manager [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 991.458766] env[62965]: DEBUG oslo_vmware.api [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390204, 'name': CopyVirtualDisk_Task} progress is 88%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.492660] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Releasing lock "refresh_cache-fe0559f0-d1ac-4af8-b554-2d31c38c2e73" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.492996] env[62965]: DEBUG nova.compute.manager [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Instance network_info: |[{"id": "75d00c8d-3b3f-4a42-b9bb-219bbf284819", "address": "fa:16:3e:1f:ce:31", "network": {"id": "c4c8235c-333c-40e7-8744-110eaa0aa23c", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1256649418-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d2d71b59ede430f846520bf49249e4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75d00c8d-3b", "ovs_interfaceid": "75d00c8d-3b3f-4a42-b9bb-219bbf284819", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 991.493659] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1f:ce:31', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dc6d5964-1106-4345-a26d-185dabd4ff0f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '75d00c8d-3b3f-4a42-b9bb-219bbf284819', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 991.503357] env[62965]: DEBUG oslo.service.loopingcall [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 991.503809] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 991.504185] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b328d71f-ef51-4117-863b-48445e8b7f5c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.536717] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 991.536717] env[62965]: value = "task-1390206" [ 991.536717] env[62965]: _type = "Task" [ 991.536717] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.552087] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.552087] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390206, 'name': CreateVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.686033] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390205, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.798204] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2685568-6902-4b42-948b-60bb60c34ee2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.817355] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Updating instance '464a3a31-06bf-4662-a5b1-d185d9a64f0b' progress to 0 {{(pid=62965) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 991.822573] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c64ca1e4-0665-457d-8ff2-c1c1116cd4d5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.830917] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3163adc1-19be-494d-9315-5cc2ddcb1fc9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.876620] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ee25ebd-9557-49f9-9a3d-70d2c26db97a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.889119] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a72a42a-959e-4bd6-bb77-3c24cf73ea20 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.907151] env[62965]: DEBUG nova.compute.provider_tree [None req-7e1a1167-a069-450d-bb6c-410b0d91a5f1 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 991.909497] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.957780] env[62965]: DEBUG oslo_vmware.api [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390204, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.251335} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.958105] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c1230705-5bdc-4c15-a740-13db3313def2/c1230705-5bdc-4c15-a740-13db3313def2.vmdk to [datastore2] a0e9cc1a-bc95-4a44-99c5-aabf85a373f4/a0e9cc1a-bc95-4a44-99c5-aabf85a373f4.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 991.958996] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52084ab7-72d3-4082-9a97-c52c93ea0613 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.985992] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Reconfiguring VM instance instance-0000004a to attach disk [datastore2] a0e9cc1a-bc95-4a44-99c5-aabf85a373f4/a0e9cc1a-bc95-4a44-99c5-aabf85a373f4.vmdk or device None with type streamOptimized {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 991.986789] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b11da2a3-7c20-40ff-927f-e3fe5ac362e2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.009523] env[62965]: DEBUG oslo_vmware.api [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 992.009523] env[62965]: value = "task-1390207" [ 992.009523] env[62965]: _type = "Task" [ 992.009523] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.018887] env[62965]: DEBUG oslo_vmware.api [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390207, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.047804] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390206, 'name': CreateVM_Task} progress is 25%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.183800] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390205, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.952068} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.184202] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 43db5f53-a719-4384-a8d2-63721dde0187/43db5f53-a719-4384-a8d2-63721dde0187.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 992.184434] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 992.184586] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b0e11eb7-0621-4b5a-88a6-6d982b881f26 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.191333] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 992.191333] env[62965]: value = "task-1390208" [ 992.191333] env[62965]: _type = "Task" [ 992.191333] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.201372] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390208, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.328277] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 992.328733] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e7dba312-0e26-4e3a-9dbc-fdf3e7eae930 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.335996] env[62965]: DEBUG oslo_vmware.api [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 992.335996] env[62965]: value = "task-1390209" [ 992.335996] env[62965]: _type = "Task" [ 992.335996] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.345020] env[62965]: DEBUG oslo_vmware.api [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390209, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.411485] env[62965]: DEBUG nova.scheduler.client.report [None req-7e1a1167-a069-450d-bb6c-410b0d91a5f1 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 992.521099] env[62965]: DEBUG oslo_vmware.api [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390207, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.549594] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390206, 'name': CreateVM_Task, 'duration_secs': 0.672885} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.549781] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 992.550525] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 992.550705] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.551057] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 992.551349] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8bccc03f-6bcf-45e0-99d2-6dc794def899 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.556271] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 992.556271] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5259c2d8-3ab7-f716-3350-a1ed81d4ee45" [ 992.556271] env[62965]: _type = "Task" [ 992.556271] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.565230] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5259c2d8-3ab7-f716-3350-a1ed81d4ee45, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.701185] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390208, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066762} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.701462] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 992.702285] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2e268e9-a47c-41cc-a9dd-560db169e25e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.725814] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] 43db5f53-a719-4384-a8d2-63721dde0187/43db5f53-a719-4384-a8d2-63721dde0187.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 992.725814] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6cf6b8e0-694e-45df-9d9b-d18f4168511e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.745039] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 992.745039] env[62965]: value = "task-1390210" [ 992.745039] env[62965]: _type = "Task" [ 992.745039] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.753635] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390210, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.846326] env[62965]: DEBUG oslo_vmware.api [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390209, 'name': PowerOffVM_Task, 'duration_secs': 0.243975} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.847415] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 992.847936] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Updating instance '464a3a31-06bf-4662-a5b1-d185d9a64f0b' progress to 17 {{(pid=62965) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 992.917120] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7e1a1167-a069-450d-bb6c-410b0d91a5f1 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.820s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.919793] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.369s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.920089] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.920264] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62965) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 992.920997] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.011s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.922489] env[62965]: INFO nova.compute.claims [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 992.927028] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da29f1ca-2679-4fe2-b977-c584f765f965 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.936478] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7463a201-c134-4e6d-8468-05c93155dba3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.944040] env[62965]: INFO nova.scheduler.client.report [None req-7e1a1167-a069-450d-bb6c-410b0d91a5f1 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Deleted allocations for instance 8a54b436-3b43-439e-bf6d-417db1fa66b2 [ 992.956590] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94260e2f-34eb-4c38-8bbc-a8e65eedf1c0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.964454] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e28cf706-e0e0-438b-99aa-0b60334f6b8e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.001931] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180018MB free_disk=174GB free_vcpus=48 pci_devices=None {{(pid=62965) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 993.002653] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.021271] env[62965]: DEBUG oslo_vmware.api [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390207, 'name': ReconfigVM_Task, 'duration_secs': 0.589203} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.021588] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Reconfigured VM instance instance-0000004a to attach disk [datastore2] a0e9cc1a-bc95-4a44-99c5-aabf85a373f4/a0e9cc1a-bc95-4a44-99c5-aabf85a373f4.vmdk or device None with type streamOptimized {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 993.022350] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-39a92a71-c5ae-404c-b390-1f738582d0d9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.029980] env[62965]: DEBUG oslo_vmware.api [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 993.029980] env[62965]: value = "task-1390211" [ 993.029980] env[62965]: _type = "Task" [ 993.029980] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.039221] env[62965]: DEBUG oslo_vmware.api [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390211, 'name': Rename_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.066977] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5259c2d8-3ab7-f716-3350-a1ed81d4ee45, 'name': SearchDatastore_Task, 'duration_secs': 0.01176} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.068110] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.068110] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 993.068110] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.068110] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.068542] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 993.068542] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5f332960-941b-462e-a560-81a76ec5efb7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.077230] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 993.077230] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 993.077923] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b33d7ab0-f477-414d-acd8-a943e588988f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.083430] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 993.083430] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52a70a2d-06e6-51d4-020b-58253e884e4c" [ 993.083430] env[62965]: _type = "Task" [ 993.083430] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.091605] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52a70a2d-06e6-51d4-020b-58253e884e4c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.258264] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390210, 'name': ReconfigVM_Task, 'duration_secs': 0.295615} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.258264] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Reconfigured VM instance instance-0000005d to attach disk [datastore2] 43db5f53-a719-4384-a8d2-63721dde0187/43db5f53-a719-4384-a8d2-63721dde0187.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 993.258264] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8a39e1f3-6611-4aa0-896f-4f2300a55c85 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.264833] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 993.264833] env[62965]: value = "task-1390212" [ 993.264833] env[62965]: _type = "Task" [ 993.264833] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.278138] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390212, 'name': Rename_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.355435] env[62965]: DEBUG nova.virt.hardware [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:24Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 993.355589] env[62965]: DEBUG nova.virt.hardware [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 993.355653] env[62965]: DEBUG nova.virt.hardware [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 993.355813] env[62965]: DEBUG nova.virt.hardware [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 993.356150] env[62965]: DEBUG nova.virt.hardware [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 993.358115] env[62965]: DEBUG nova.virt.hardware [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 993.358115] env[62965]: DEBUG nova.virt.hardware [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 993.358115] env[62965]: DEBUG nova.virt.hardware [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 993.358115] env[62965]: DEBUG nova.virt.hardware [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 993.358115] env[62965]: DEBUG nova.virt.hardware [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 993.358115] env[62965]: DEBUG nova.virt.hardware [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 993.362945] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4ae23305-2f2f-4d5a-bd6e-a132cf2f11d2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.380215] env[62965]: DEBUG oslo_vmware.api [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 993.380215] env[62965]: value = "task-1390213" [ 993.380215] env[62965]: _type = "Task" [ 993.380215] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.389677] env[62965]: DEBUG oslo_vmware.api [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390213, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.466627] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7e1a1167-a069-450d-bb6c-410b0d91a5f1 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "8a54b436-3b43-439e-bf6d-417db1fa66b2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.982s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.542327] env[62965]: DEBUG oslo_vmware.api [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390211, 'name': Rename_Task, 'duration_secs': 0.135284} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.542327] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 993.542563] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ce920689-f526-481a-babb-123b0367b14f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.549527] env[62965]: DEBUG oslo_vmware.api [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 993.549527] env[62965]: value = "task-1390214" [ 993.549527] env[62965]: _type = "Task" [ 993.549527] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.561951] env[62965]: DEBUG oslo_vmware.api [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390214, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.594471] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52a70a2d-06e6-51d4-020b-58253e884e4c, 'name': SearchDatastore_Task, 'duration_secs': 0.009577} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.596209] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2429407c-0dd3-42f9-87bb-0690b485c784 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.601916] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 993.601916] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5220d14b-e9ad-1cbc-69e5-0ccf435b9a30" [ 993.601916] env[62965]: _type = "Task" [ 993.601916] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.611115] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5220d14b-e9ad-1cbc-69e5-0ccf435b9a30, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.775459] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390212, 'name': Rename_Task, 'duration_secs': 0.143584} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.775459] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 993.775459] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c6fc3003-d9ae-4a1f-b421-3600d33fba86 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.781933] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 993.781933] env[62965]: value = "task-1390215" [ 993.781933] env[62965]: _type = "Task" [ 993.781933] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.790112] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390215, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.889968] env[62965]: DEBUG oslo_vmware.api [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390213, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.059706] env[62965]: DEBUG oslo_vmware.api [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390214, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.111746] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5220d14b-e9ad-1cbc-69e5-0ccf435b9a30, 'name': SearchDatastore_Task, 'duration_secs': 0.010037} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.114513] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.114792] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] fe0559f0-d1ac-4af8-b554-2d31c38c2e73/fe0559f0-d1ac-4af8-b554-2d31c38c2e73.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 994.115261] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bf0615ef-4c20-46b9-9a5b-9102206cf21a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.122028] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 994.122028] env[62965]: value = "task-1390216" [ 994.122028] env[62965]: _type = "Task" [ 994.122028] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.127337] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b81a38d9-568a-4719-aa4e-db7235f9606c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.134601] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390216, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.137337] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-239f6857-07e3-4eef-96db-9711a3a0d1cc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.172193] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-978afbfa-dcc9-44b8-aae6-28347a939238 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.179852] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7e1c066-543c-4250-b355-015c11013acc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.193663] env[62965]: DEBUG nova.compute.provider_tree [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 994.210192] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8ba88bcb-0aeb-416d-99c7-a16d2ed15b4f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "08f7b05f-3fb5-48ac-a68a-1d6fd804a622" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.210543] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8ba88bcb-0aeb-416d-99c7-a16d2ed15b4f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "08f7b05f-3fb5-48ac-a68a-1d6fd804a622" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.211056] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8ba88bcb-0aeb-416d-99c7-a16d2ed15b4f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "08f7b05f-3fb5-48ac-a68a-1d6fd804a622-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.211056] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8ba88bcb-0aeb-416d-99c7-a16d2ed15b4f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "08f7b05f-3fb5-48ac-a68a-1d6fd804a622-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.211209] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8ba88bcb-0aeb-416d-99c7-a16d2ed15b4f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "08f7b05f-3fb5-48ac-a68a-1d6fd804a622-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.213334] env[62965]: INFO nova.compute.manager [None req-8ba88bcb-0aeb-416d-99c7-a16d2ed15b4f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Terminating instance [ 994.250767] env[62965]: DEBUG oslo_vmware.rw_handles [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5209d8fd-b8a4-6232-3d48-f3099d427df7/disk-0.vmdk. {{(pid=62965) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 994.251782] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daf5a8da-411b-4e85-bc76-3c0ae6fb6820 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.258120] env[62965]: DEBUG oslo_vmware.rw_handles [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5209d8fd-b8a4-6232-3d48-f3099d427df7/disk-0.vmdk is in state: ready. {{(pid=62965) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 994.258341] env[62965]: ERROR oslo_vmware.rw_handles [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5209d8fd-b8a4-6232-3d48-f3099d427df7/disk-0.vmdk due to incomplete transfer. [ 994.258591] env[62965]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-b373d5e7-1e31-4c98-ac6f-bb325264db14 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.266368] env[62965]: DEBUG oslo_vmware.rw_handles [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5209d8fd-b8a4-6232-3d48-f3099d427df7/disk-0.vmdk. {{(pid=62965) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 994.266566] env[62965]: DEBUG nova.virt.vmwareapi.images [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Uploaded image a792f1cf-fa3a-46b4-9178-12b6f8866058 to the Glance image server {{(pid=62965) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 994.269042] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Destroying the VM {{(pid=62965) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 994.269245] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-5acd2a6f-9d6f-47b1-95fb-3a0c09316b36 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.274892] env[62965]: DEBUG oslo_vmware.api [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 994.274892] env[62965]: value = "task-1390217" [ 994.274892] env[62965]: _type = "Task" [ 994.274892] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.283059] env[62965]: DEBUG oslo_vmware.api [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390217, 'name': Destroy_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.291440] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390215, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.392176] env[62965]: DEBUG oslo_vmware.api [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390213, 'name': ReconfigVM_Task, 'duration_secs': 0.933192} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.392431] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Updating instance '464a3a31-06bf-4662-a5b1-d185d9a64f0b' progress to 33 {{(pid=62965) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 994.562681] env[62965]: DEBUG oslo_vmware.api [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390214, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.633184] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390216, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.696606] env[62965]: DEBUG nova.scheduler.client.report [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 994.717423] env[62965]: DEBUG nova.compute.manager [None req-8ba88bcb-0aeb-416d-99c7-a16d2ed15b4f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 994.717751] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-8ba88bcb-0aeb-416d-99c7-a16d2ed15b4f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 994.719017] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc989f0f-3b50-4751-a152-7268724f548c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.727156] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ba88bcb-0aeb-416d-99c7-a16d2ed15b4f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 994.727512] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9c0b661d-a8e8-41ed-ad0f-4b9cc62ebc71 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.732604] env[62965]: DEBUG oslo_vmware.api [None req-8ba88bcb-0aeb-416d-99c7-a16d2ed15b4f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 994.732604] env[62965]: value = "task-1390218" [ 994.732604] env[62965]: _type = "Task" [ 994.732604] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.740192] env[62965]: DEBUG oslo_vmware.api [None req-8ba88bcb-0aeb-416d-99c7-a16d2ed15b4f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390218, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.784275] env[62965]: DEBUG oslo_vmware.api [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390217, 'name': Destroy_Task, 'duration_secs': 0.366023} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.787403] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Destroyed the VM [ 994.787658] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Deleting Snapshot of the VM instance {{(pid=62965) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 994.788233] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-3f2a8e19-5efa-4174-a7c6-13e47f9bad4a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.794528] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390215, 'name': PowerOnVM_Task, 'duration_secs': 0.641722} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.795683] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 994.795898] env[62965]: INFO nova.compute.manager [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Took 10.07 seconds to spawn the instance on the hypervisor. [ 994.796100] env[62965]: DEBUG nova.compute.manager [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 994.796411] env[62965]: DEBUG oslo_vmware.api [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 994.796411] env[62965]: value = "task-1390219" [ 994.796411] env[62965]: _type = "Task" [ 994.796411] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.797130] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dd2b04c-35bb-4312-aeeb-f81779532278 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.812697] env[62965]: DEBUG oslo_vmware.api [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390219, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.899277] env[62965]: DEBUG nova.virt.hardware [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 994.899532] env[62965]: DEBUG nova.virt.hardware [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 994.899702] env[62965]: DEBUG nova.virt.hardware [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 994.899898] env[62965]: DEBUG nova.virt.hardware [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 994.900097] env[62965]: DEBUG nova.virt.hardware [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 994.900259] env[62965]: DEBUG nova.virt.hardware [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 994.900537] env[62965]: DEBUG nova.virt.hardware [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 994.900707] env[62965]: DEBUG nova.virt.hardware [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 994.900876] env[62965]: DEBUG nova.virt.hardware [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 994.901051] env[62965]: DEBUG nova.virt.hardware [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 994.901230] env[62965]: DEBUG nova.virt.hardware [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 994.906476] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Reconfiguring VM instance instance-0000005a to detach disk 2000 {{(pid=62965) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 994.906763] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f51da876-af37-445c-9345-3512245dcb5d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.925238] env[62965]: DEBUG oslo_vmware.api [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 994.925238] env[62965]: value = "task-1390220" [ 994.925238] env[62965]: _type = "Task" [ 994.925238] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.933057] env[62965]: DEBUG oslo_vmware.api [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390220, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.061053] env[62965]: DEBUG oslo_vmware.api [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390214, 'name': PowerOnVM_Task, 'duration_secs': 1.153312} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.061238] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 995.133622] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390216, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.631021} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.133905] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] fe0559f0-d1ac-4af8-b554-2d31c38c2e73/fe0559f0-d1ac-4af8-b554-2d31c38c2e73.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 995.134146] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 995.134505] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cf697995-3e44-49c0-80cc-fbf112cf15dd {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.141103] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 995.141103] env[62965]: value = "task-1390221" [ 995.141103] env[62965]: _type = "Task" [ 995.141103] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.149177] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390221, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.164380] env[62965]: DEBUG nova.compute.manager [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 995.165455] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d445b036-91bd-4cb7-9916-e5b1459dd3f5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.202840] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.282s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.203359] env[62965]: DEBUG nova.compute.manager [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 995.205864] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 2.204s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.243174] env[62965]: DEBUG oslo_vmware.api [None req-8ba88bcb-0aeb-416d-99c7-a16d2ed15b4f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390218, 'name': PowerOffVM_Task, 'duration_secs': 0.366571} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.243470] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ba88bcb-0aeb-416d-99c7-a16d2ed15b4f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 995.243671] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-8ba88bcb-0aeb-416d-99c7-a16d2ed15b4f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 995.243926] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-46ae0b99-14b6-437d-b046-10b2b6e7bc34 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.328711] env[62965]: INFO nova.compute.manager [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Took 22.39 seconds to build instance. [ 995.333624] env[62965]: DEBUG oslo_vmware.api [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390219, 'name': RemoveSnapshot_Task, 'duration_secs': 0.378629} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.334878] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Deleted Snapshot of the VM instance {{(pid=62965) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 995.335474] env[62965]: DEBUG nova.compute.manager [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 995.335474] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-8ba88bcb-0aeb-416d-99c7-a16d2ed15b4f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 995.335637] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-8ba88bcb-0aeb-416d-99c7-a16d2ed15b4f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Deleting contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 995.335802] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ba88bcb-0aeb-416d-99c7-a16d2ed15b4f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Deleting the datastore file [datastore2] 08f7b05f-3fb5-48ac-a68a-1d6fd804a622 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 995.336685] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1034ae3-c6fd-4096-a910-4821c3b77ef3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.339195] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9f5ac3ad-92e2-4703-a43f-1ff634351f5d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.349358] env[62965]: DEBUG oslo_vmware.api [None req-8ba88bcb-0aeb-416d-99c7-a16d2ed15b4f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 995.349358] env[62965]: value = "task-1390223" [ 995.349358] env[62965]: _type = "Task" [ 995.349358] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.357398] env[62965]: DEBUG oslo_vmware.api [None req-8ba88bcb-0aeb-416d-99c7-a16d2ed15b4f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390223, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.435742] env[62965]: DEBUG oslo_vmware.api [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390220, 'name': ReconfigVM_Task, 'duration_secs': 0.30174} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.436504] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Reconfigured VM instance instance-0000005a to detach disk 2000 {{(pid=62965) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 995.437397] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb55e9ec-b19d-41ee-99d5-0d1bd052c4a5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.461884] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] 464a3a31-06bf-4662-a5b1-d185d9a64f0b/464a3a31-06bf-4662-a5b1-d185d9a64f0b.vmdk or device None with type thin {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 995.462215] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f24d8f10-9ceb-49fe-83b9-18d170f696cb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.493124] env[62965]: DEBUG oslo_vmware.api [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 995.493124] env[62965]: value = "task-1390224" [ 995.493124] env[62965]: _type = "Task" [ 995.493124] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.502396] env[62965]: DEBUG oslo_vmware.api [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390224, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.651047] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390221, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074538} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.651047] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 995.651577] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87d82f16-c018-45b3-96c4-c1e18d371285 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.674561] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] fe0559f0-d1ac-4af8-b554-2d31c38c2e73/fe0559f0-d1ac-4af8-b554-2d31c38c2e73.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 995.674881] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4b34a8a0-ca32-4ea1-8172-b4e9ad4d3afa {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.697833] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 995.697833] env[62965]: value = "task-1390225" [ 995.697833] env[62965]: _type = "Task" [ 995.697833] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.698303] env[62965]: DEBUG oslo_concurrency.lockutils [None req-348661f3-c0fa-48cb-a913-ea21cdc0afc2 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lock "a0e9cc1a-bc95-4a44-99c5-aabf85a373f4" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 25.239s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.707017] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390225, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.710235] env[62965]: DEBUG nova.compute.utils [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 995.722665] env[62965]: DEBUG nova.compute.manager [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 995.723939] env[62965]: DEBUG nova.network.neutron [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 995.762334] env[62965]: DEBUG nova.policy [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e9d9f79c936143f5815fcb99755bbd4e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8fab418e2fe1420793517663574b43bb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 995.830800] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "43db5f53-a719-4384-a8d2-63721dde0187" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.905s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.855199] env[62965]: INFO nova.compute.manager [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Shelve offloading [ 995.861777] env[62965]: DEBUG oslo_vmware.api [None req-8ba88bcb-0aeb-416d-99c7-a16d2ed15b4f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390223, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.352697} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.862013] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ba88bcb-0aeb-416d-99c7-a16d2ed15b4f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 995.862558] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-8ba88bcb-0aeb-416d-99c7-a16d2ed15b4f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Deleted contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 995.862822] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-8ba88bcb-0aeb-416d-99c7-a16d2ed15b4f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 995.863053] env[62965]: INFO nova.compute.manager [None req-8ba88bcb-0aeb-416d-99c7-a16d2ed15b4f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Took 1.15 seconds to destroy the instance on the hypervisor. [ 995.863321] env[62965]: DEBUG oslo.service.loopingcall [None req-8ba88bcb-0aeb-416d-99c7-a16d2ed15b4f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 995.863531] env[62965]: DEBUG nova.compute.manager [-] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 995.863621] env[62965]: DEBUG nova.network.neutron [-] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 996.000978] env[62965]: DEBUG oslo_vmware.api [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390224, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.018827] env[62965]: DEBUG nova.network.neutron [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Successfully created port: ee7ad38e-9936-45bf-8e20-b20e5f7958bd {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 996.209546] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390225, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.222879] env[62965]: DEBUG nova.compute.manager [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 996.227529] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Applying migration context for instance 464a3a31-06bf-4662-a5b1-d185d9a64f0b as it has an incoming, in-progress migration a802f9c1-2aa9-4613-9e17-47bbb76297fb. Migration status is migrating {{(pid=62965) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 996.227529] env[62965]: INFO nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Updating resource usage from migration a802f9c1-2aa9-4613-9e17-47bbb76297fb [ 996.242218] env[62965]: DEBUG nova.compute.manager [req-08f8cf39-41a8-46e1-a458-db7ae72b8417 req-38ac130c-9c95-4b4a-9378-64063c0c4541 service nova] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Received event network-vif-deleted-6b5a028e-baa2-4d3d-a9ca-ade1c685e4b9 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 996.242218] env[62965]: INFO nova.compute.manager [req-08f8cf39-41a8-46e1-a458-db7ae72b8417 req-38ac130c-9c95-4b4a-9378-64063c0c4541 service nova] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Neutron deleted interface 6b5a028e-baa2-4d3d-a9ca-ade1c685e4b9; detaching it from the instance and deleting it from the info cache [ 996.242218] env[62965]: DEBUG nova.network.neutron [req-08f8cf39-41a8-46e1-a458-db7ae72b8417 req-38ac130c-9c95-4b4a-9378-64063c0c4541 service nova] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.254353] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance c2448b76-1553-4b68-a731-0bb0967f4c1d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 996.254353] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 924c5cdc-0450-4ced-b920-50e2d5060fd2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 996.254353] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance bd36601b-5a70-4a29-8ef1-d1e925f41de7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 996.254353] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance a63b301d-8d52-4fec-951f-a15cda9449ac actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 996.254353] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance b59d6023-a047-4b32-bd18-1c7cc05f67d7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 996.254353] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 08f7b05f-3fb5-48ac-a68a-1d6fd804a622 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 996.254353] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 51babf43-b7b0-4731-a989-6977874fb801 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 996.254353] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance a0e9cc1a-bc95-4a44-99c5-aabf85a373f4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 996.254353] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 43db5f53-a719-4384-a8d2-63721dde0187 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 996.254353] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance fe0559f0-d1ac-4af8-b554-2d31c38c2e73 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 996.254353] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Migration a802f9c1-2aa9-4613-9e17-47bbb76297fb is active on this compute host and has allocations in placement: {'resources': {'VCPU': 1, 'MEMORY_MB': 192, 'DISK_GB': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 996.254353] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 464a3a31-06bf-4662-a5b1-d185d9a64f0b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 996.254353] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 8efad776-6cbe-49ab-83dc-a410d8417505 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 996.254353] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Total usable vcpus: 48, total allocated vcpus: 13 {{(pid=62965) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 996.254887] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3072MB phys_disk=200GB used_disk=13GB total_vcpus=48 used_vcpus=13 pci_stats=[] {{(pid=62965) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 996.358594] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 996.358903] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-929cdf9e-d28b-4424-b077-23b227a4f2e9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.366117] env[62965]: DEBUG oslo_vmware.api [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 996.366117] env[62965]: value = "task-1390226" [ 996.366117] env[62965]: _type = "Task" [ 996.366117] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.376640] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] VM already powered off {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 996.376912] env[62965]: DEBUG nova.compute.manager [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 996.377670] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f397dd2-d64b-4cfd-8c46-22d98bce3c14 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.385419] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "refresh_cache-51babf43-b7b0-4731-a989-6977874fb801" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.385590] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquired lock "refresh_cache-51babf43-b7b0-4731-a989-6977874fb801" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.385762] env[62965]: DEBUG nova.network.neutron [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 996.426815] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59a762e8-dcd3-4df0-bdb2-6a62f8023120 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.434754] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26165a5a-c769-4444-85f2-9fc396023e3c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.466318] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-109e77f0-30dd-4853-87db-c3443d70924f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.474038] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3126b93-5cbf-477c-8667-04c7fa4f4fdd {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.487620] env[62965]: DEBUG nova.compute.provider_tree [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 996.499840] env[62965]: DEBUG oslo_vmware.api [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390224, 'name': ReconfigVM_Task, 'duration_secs': 0.845606} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.500381] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Reconfigured VM instance instance-0000005a to attach disk [datastore1] 464a3a31-06bf-4662-a5b1-d185d9a64f0b/464a3a31-06bf-4662-a5b1-d185d9a64f0b.vmdk or device None with type thin {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 996.500685] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Updating instance '464a3a31-06bf-4662-a5b1-d185d9a64f0b' progress to 50 {{(pid=62965) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 996.711380] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390225, 'name': ReconfigVM_Task, 'duration_secs': 0.796507} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.711745] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Reconfigured VM instance instance-0000005e to attach disk [datastore2] fe0559f0-d1ac-4af8-b554-2d31c38c2e73/fe0559f0-d1ac-4af8-b554-2d31c38c2e73.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 996.712382] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4ec04acb-4f16-4a90-94a3-5ae492437940 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.713922] env[62965]: DEBUG nova.network.neutron [-] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.720452] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 996.720452] env[62965]: value = "task-1390227" [ 996.720452] env[62965]: _type = "Task" [ 996.720452] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.734299] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390227, 'name': Rename_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.746637] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c109af4f-6578-4805-92a1-a4bb399bcdec {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.755627] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29f6e2eb-9a3b-4a5a-ba1b-acca39cca99e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.786176] env[62965]: DEBUG nova.compute.manager [req-08f8cf39-41a8-46e1-a458-db7ae72b8417 req-38ac130c-9c95-4b4a-9378-64063c0c4541 service nova] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Detach interface failed, port_id=6b5a028e-baa2-4d3d-a9ca-ade1c685e4b9, reason: Instance 08f7b05f-3fb5-48ac-a68a-1d6fd804a622 could not be found. {{(pid=62965) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11294}} [ 996.990884] env[62965]: DEBUG nova.scheduler.client.report [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 997.007059] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c020d46-cdbf-4138-b5a7-2d8b7963f4d2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.030213] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7591be4-6f73-47d9-879e-872634831f02 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.050736] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Updating instance '464a3a31-06bf-4662-a5b1-d185d9a64f0b' progress to 67 {{(pid=62965) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 997.153550] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b1d08d7-5863-440b-9180-d0c3b529bb3c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.160774] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-5080be0c-34fd-491e-98cd-5e7f3f048929 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Suspending the VM {{(pid=62965) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 997.161022] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-52548253-e024-4409-ad71-f0a775b153af {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.167396] env[62965]: DEBUG oslo_vmware.api [None req-5080be0c-34fd-491e-98cd-5e7f3f048929 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 997.167396] env[62965]: value = "task-1390228" [ 997.167396] env[62965]: _type = "Task" [ 997.167396] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.177290] env[62965]: DEBUG oslo_vmware.api [None req-5080be0c-34fd-491e-98cd-5e7f3f048929 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390228, 'name': SuspendVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.214368] env[62965]: DEBUG nova.network.neutron [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Updating instance_info_cache with network_info: [{"id": "2664b4df-ad2e-4114-8837-6c5429ae94e3", "address": "fa:16:3e:52:04:a7", "network": {"id": "62052700-1781-4933-93c8-fa1aafdb0fd2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1931798228-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "978666785b0743bb843de96585f96fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2664b4df-ad", "ovs_interfaceid": "2664b4df-ad2e-4114-8837-6c5429ae94e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.217268] env[62965]: INFO nova.compute.manager [-] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Took 1.35 seconds to deallocate network for instance. [ 997.231046] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390227, 'name': Rename_Task, 'duration_secs': 0.251577} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.231193] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 997.231418] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-89f55df6-7c90-408c-99fc-8b25d29df943 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.236135] env[62965]: DEBUG nova.compute.manager [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 997.240453] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 997.240453] env[62965]: value = "task-1390229" [ 997.240453] env[62965]: _type = "Task" [ 997.240453] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.249960] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390229, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.262252] env[62965]: DEBUG nova.virt.hardware [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 997.262640] env[62965]: DEBUG nova.virt.hardware [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 997.262745] env[62965]: DEBUG nova.virt.hardware [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 997.262926] env[62965]: DEBUG nova.virt.hardware [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 997.263112] env[62965]: DEBUG nova.virt.hardware [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 997.263307] env[62965]: DEBUG nova.virt.hardware [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 997.263544] env[62965]: DEBUG nova.virt.hardware [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 997.263713] env[62965]: DEBUG nova.virt.hardware [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 997.263899] env[62965]: DEBUG nova.virt.hardware [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 997.264078] env[62965]: DEBUG nova.virt.hardware [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 997.264313] env[62965]: DEBUG nova.virt.hardware [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 997.265546] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5516e355-8c5b-494d-93a0-8968883c2763 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.274610] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bafc18eb-f087-4463-bc63-5e8a08b95990 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.429534] env[62965]: DEBUG nova.compute.manager [req-93d87adb-e247-49f4-9608-e3d249b98dcb req-d5456940-bb80-405c-959a-afb96d402a52 service nova] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Received event network-vif-plugged-ee7ad38e-9936-45bf-8e20-b20e5f7958bd {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 997.429534] env[62965]: DEBUG oslo_concurrency.lockutils [req-93d87adb-e247-49f4-9608-e3d249b98dcb req-d5456940-bb80-405c-959a-afb96d402a52 service nova] Acquiring lock "8efad776-6cbe-49ab-83dc-a410d8417505-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.429534] env[62965]: DEBUG oslo_concurrency.lockutils [req-93d87adb-e247-49f4-9608-e3d249b98dcb req-d5456940-bb80-405c-959a-afb96d402a52 service nova] Lock "8efad776-6cbe-49ab-83dc-a410d8417505-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.429534] env[62965]: DEBUG oslo_concurrency.lockutils [req-93d87adb-e247-49f4-9608-e3d249b98dcb req-d5456940-bb80-405c-959a-afb96d402a52 service nova] Lock "8efad776-6cbe-49ab-83dc-a410d8417505-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.429534] env[62965]: DEBUG nova.compute.manager [req-93d87adb-e247-49f4-9608-e3d249b98dcb req-d5456940-bb80-405c-959a-afb96d402a52 service nova] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] No waiting events found dispatching network-vif-plugged-ee7ad38e-9936-45bf-8e20-b20e5f7958bd {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 997.430103] env[62965]: WARNING nova.compute.manager [req-93d87adb-e247-49f4-9608-e3d249b98dcb req-d5456940-bb80-405c-959a-afb96d402a52 service nova] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Received unexpected event network-vif-plugged-ee7ad38e-9936-45bf-8e20-b20e5f7958bd for instance with vm_state building and task_state spawning. [ 997.496553] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62965) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 997.496553] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.291s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.528161] env[62965]: DEBUG nova.network.neutron [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Successfully updated port: ee7ad38e-9936-45bf-8e20-b20e5f7958bd {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 997.591965] env[62965]: DEBUG nova.network.neutron [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Port 170ea80e-f3c0-4647-af9d-34ffed538c33 binding to destination host cpu-1 is already ACTIVE {{(pid=62965) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 997.679762] env[62965]: DEBUG oslo_vmware.api [None req-5080be0c-34fd-491e-98cd-5e7f3f048929 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390228, 'name': SuspendVM_Task} progress is 62%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.717258] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Releasing lock "refresh_cache-51babf43-b7b0-4731-a989-6977874fb801" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.723685] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8ba88bcb-0aeb-416d-99c7-a16d2ed15b4f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.723685] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8ba88bcb-0aeb-416d-99c7-a16d2ed15b4f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.723858] env[62965]: DEBUG nova.objects.instance [None req-8ba88bcb-0aeb-416d-99c7-a16d2ed15b4f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lazy-loading 'resources' on Instance uuid 08f7b05f-3fb5-48ac-a68a-1d6fd804a622 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 997.752245] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390229, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.771645] env[62965]: DEBUG oslo_concurrency.lockutils [None req-06d11f1e-7c69-424b-9cfb-f1be5384bcba tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquiring lock "a63b301d-8d52-4fec-951f-a15cda9449ac" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.772015] env[62965]: DEBUG oslo_concurrency.lockutils [None req-06d11f1e-7c69-424b-9cfb-f1be5384bcba tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "a63b301d-8d52-4fec-951f-a15cda9449ac" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.993479] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 997.993479] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9539fbee-71c9-4f90-b631-5f98a533a2d0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.000854] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 998.001116] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0d362e57-8dfd-41d0-9d2d-c0a114380c0e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.031498] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "refresh_cache-8efad776-6cbe-49ab-83dc-a410d8417505" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 998.031735] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquired lock "refresh_cache-8efad776-6cbe-49ab-83dc-a410d8417505" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.031903] env[62965]: DEBUG nova.network.neutron [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 998.091817] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 998.092140] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Deleting contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 998.092418] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Deleting the datastore file [datastore1] 51babf43-b7b0-4731-a989-6977874fb801 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 998.092754] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1147c42c-94e8-4645-9288-7e1f348a922d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.108025] env[62965]: DEBUG oslo_vmware.api [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 998.108025] env[62965]: value = "task-1390231" [ 998.108025] env[62965]: _type = "Task" [ 998.108025] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.115871] env[62965]: DEBUG oslo_vmware.api [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390231, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.179212] env[62965]: DEBUG oslo_vmware.api [None req-5080be0c-34fd-491e-98cd-5e7f3f048929 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390228, 'name': SuspendVM_Task, 'duration_secs': 0.707832} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.179487] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-5080be0c-34fd-491e-98cd-5e7f3f048929 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Suspended the VM {{(pid=62965) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 998.179671] env[62965]: DEBUG nova.compute.manager [None req-5080be0c-34fd-491e-98cd-5e7f3f048929 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 998.180485] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ccacd8d-1d38-4b77-bbeb-fd3f46b708b2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.254196] env[62965]: DEBUG oslo_vmware.api [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390229, 'name': PowerOnVM_Task, 'duration_secs': 0.607217} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.254551] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 998.254797] env[62965]: INFO nova.compute.manager [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Took 10.11 seconds to spawn the instance on the hypervisor. [ 998.255039] env[62965]: DEBUG nova.compute.manager [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 998.255831] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-781e2f28-586f-4340-97c3-0a3913b806d5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.275414] env[62965]: DEBUG nova.compute.utils [None req-06d11f1e-7c69-424b-9cfb-f1be5384bcba tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 998.381957] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4b78198-159d-48c1-8982-a415a502eb7d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.390941] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ad29e20-4f38-47fe-8ddd-0fd26e724966 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.421335] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eeda0ab-c8c0-49b8-bb08-9b83aeeef554 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.429436] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c3c8b38-3f1d-4ce7-ae61-cb3e85946972 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.445204] env[62965]: DEBUG nova.compute.provider_tree [None req-8ba88bcb-0aeb-416d-99c7-a16d2ed15b4f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 998.629629] env[62965]: DEBUG oslo_concurrency.lockutils [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "464a3a31-06bf-4662-a5b1-d185d9a64f0b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.629880] env[62965]: DEBUG oslo_concurrency.lockutils [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "464a3a31-06bf-4662-a5b1-d185d9a64f0b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.630497] env[62965]: DEBUG oslo_concurrency.lockutils [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "464a3a31-06bf-4662-a5b1-d185d9a64f0b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.631475] env[62965]: DEBUG nova.network.neutron [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 998.639485] env[62965]: DEBUG oslo_vmware.api [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390231, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.760417] env[62965]: DEBUG nova.network.neutron [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Updating instance_info_cache with network_info: [{"id": "ee7ad38e-9936-45bf-8e20-b20e5f7958bd", "address": "fa:16:3e:93:d2:25", "network": {"id": "4649fed7-c2ca-43a9-a183-7c99423da726", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1978698173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fab418e2fe1420793517663574b43bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43ad01d2-c7dd-453c-a929-8ad76294d13c", "external-id": "nsx-vlan-transportzone-176", "segmentation_id": 176, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee7ad38e-99", "ovs_interfaceid": "ee7ad38e-9936-45bf-8e20-b20e5f7958bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.775701] env[62965]: INFO nova.compute.manager [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Took 25.81 seconds to build instance. [ 998.778173] env[62965]: DEBUG oslo_concurrency.lockutils [None req-06d11f1e-7c69-424b-9cfb-f1be5384bcba tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "a63b301d-8d52-4fec-951f-a15cda9449ac" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.948825] env[62965]: DEBUG nova.scheduler.client.report [None req-8ba88bcb-0aeb-416d-99c7-a16d2ed15b4f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 999.125326] env[62965]: DEBUG oslo_vmware.api [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390231, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.601805} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.125578] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 999.125780] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Deleted contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 999.125982] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 999.144100] env[62965]: INFO nova.scheduler.client.report [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Deleted allocations for instance 51babf43-b7b0-4731-a989-6977874fb801 [ 999.263441] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Releasing lock "refresh_cache-8efad776-6cbe-49ab-83dc-a410d8417505" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.263804] env[62965]: DEBUG nova.compute.manager [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Instance network_info: |[{"id": "ee7ad38e-9936-45bf-8e20-b20e5f7958bd", "address": "fa:16:3e:93:d2:25", "network": {"id": "4649fed7-c2ca-43a9-a183-7c99423da726", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1978698173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fab418e2fe1420793517663574b43bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43ad01d2-c7dd-453c-a929-8ad76294d13c", "external-id": "nsx-vlan-transportzone-176", "segmentation_id": 176, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee7ad38e-99", "ovs_interfaceid": "ee7ad38e-9936-45bf-8e20-b20e5f7958bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 999.264252] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:93:d2:25', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '43ad01d2-c7dd-453c-a929-8ad76294d13c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ee7ad38e-9936-45bf-8e20-b20e5f7958bd', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 999.271774] env[62965]: DEBUG oslo.service.loopingcall [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 999.271980] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 999.272219] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6067582e-4e80-45e3-b5f4-984dbf8c3677 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.286708] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3890eed8-d242-42e8-a9c8-f6547b2350b7 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "fe0559f0-d1ac-4af8-b554-2d31c38c2e73" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.330s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.292828] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 999.292828] env[62965]: value = "task-1390232" [ 999.292828] env[62965]: _type = "Task" [ 999.292828] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.300527] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390232, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.453855] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8ba88bcb-0aeb-416d-99c7-a16d2ed15b4f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.730s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.472338] env[62965]: INFO nova.scheduler.client.report [None req-8ba88bcb-0aeb-416d-99c7-a16d2ed15b4f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Deleted allocations for instance 08f7b05f-3fb5-48ac-a68a-1d6fd804a622 [ 999.539511] env[62965]: DEBUG nova.compute.manager [req-60f0d1db-05b5-4302-b625-ab15d8eec802 req-3011a1b5-5b89-4de7-b837-4ca1c1dfc3bd service nova] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Received event network-changed-ee7ad38e-9936-45bf-8e20-b20e5f7958bd {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 999.539750] env[62965]: DEBUG nova.compute.manager [req-60f0d1db-05b5-4302-b625-ab15d8eec802 req-3011a1b5-5b89-4de7-b837-4ca1c1dfc3bd service nova] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Refreshing instance network info cache due to event network-changed-ee7ad38e-9936-45bf-8e20-b20e5f7958bd. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 999.539978] env[62965]: DEBUG oslo_concurrency.lockutils [req-60f0d1db-05b5-4302-b625-ab15d8eec802 req-3011a1b5-5b89-4de7-b837-4ca1c1dfc3bd service nova] Acquiring lock "refresh_cache-8efad776-6cbe-49ab-83dc-a410d8417505" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.540139] env[62965]: DEBUG oslo_concurrency.lockutils [req-60f0d1db-05b5-4302-b625-ab15d8eec802 req-3011a1b5-5b89-4de7-b837-4ca1c1dfc3bd service nova] Acquired lock "refresh_cache-8efad776-6cbe-49ab-83dc-a410d8417505" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.540306] env[62965]: DEBUG nova.network.neutron [req-60f0d1db-05b5-4302-b625-ab15d8eec802 req-3011a1b5-5b89-4de7-b837-4ca1c1dfc3bd service nova] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Refreshing network info cache for port ee7ad38e-9936-45bf-8e20-b20e5f7958bd {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 999.648769] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.649049] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.649282] env[62965]: DEBUG nova.objects.instance [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lazy-loading 'resources' on Instance uuid 51babf43-b7b0-4731-a989-6977874fb801 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 999.666933] env[62965]: DEBUG oslo_concurrency.lockutils [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "refresh_cache-464a3a31-06bf-4662-a5b1-d185d9a64f0b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.667135] env[62965]: DEBUG oslo_concurrency.lockutils [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquired lock "refresh_cache-464a3a31-06bf-4662-a5b1-d185d9a64f0b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.667311] env[62965]: DEBUG nova.network.neutron [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 999.718075] env[62965]: INFO nova.compute.manager [None req-500a830e-d2d3-465f-9a8b-456fd01cb73a tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Resuming [ 999.718686] env[62965]: DEBUG nova.objects.instance [None req-500a830e-d2d3-465f-9a8b-456fd01cb73a tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lazy-loading 'flavor' on Instance uuid a0e9cc1a-bc95-4a44-99c5-aabf85a373f4 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 999.802739] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390232, 'name': CreateVM_Task, 'duration_secs': 0.383757} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.803028] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 999.803578] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.803872] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.804242] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 999.804514] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-21f69408-5114-4b73-98d8-eccd5da8b817 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.808599] env[62965]: DEBUG oslo_vmware.api [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 999.808599] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52c92024-6d43-c793-b5f1-f592f9f945bf" [ 999.808599] env[62965]: _type = "Task" [ 999.808599] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.815753] env[62965]: DEBUG oslo_vmware.api [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52c92024-6d43-c793-b5f1-f592f9f945bf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.842884] env[62965]: DEBUG oslo_concurrency.lockutils [None req-06d11f1e-7c69-424b-9cfb-f1be5384bcba tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquiring lock "a63b301d-8d52-4fec-951f-a15cda9449ac" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.843125] env[62965]: DEBUG oslo_concurrency.lockutils [None req-06d11f1e-7c69-424b-9cfb-f1be5384bcba tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "a63b301d-8d52-4fec-951f-a15cda9449ac" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.843358] env[62965]: INFO nova.compute.manager [None req-06d11f1e-7c69-424b-9cfb-f1be5384bcba tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Attaching volume bbfeace6-81aa-40a3-8563-95c2c9a1c7ef to /dev/sdb [ 999.878601] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a7ee55d-6de5-460c-95ba-314ce37eebf2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.885389] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ca02c7e-2589-4815-9d3d-be4bb6aae641 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.898634] env[62965]: DEBUG nova.virt.block_device [None req-06d11f1e-7c69-424b-9cfb-f1be5384bcba tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Updating existing volume attachment record: cd1bee05-8895-4107-9181-cc1267283666 {{(pid=62965) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 999.984370] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8ba88bcb-0aeb-416d-99c7-a16d2ed15b4f tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "08f7b05f-3fb5-48ac-a68a-1d6fd804a622" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.774s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.132309] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d866e5f5-3d83-4c03-a70d-9af52ea25867 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquiring lock "43db5f53-a719-4384-a8d2-63721dde0187" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.132567] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d866e5f5-3d83-4c03-a70d-9af52ea25867 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "43db5f53-a719-4384-a8d2-63721dde0187" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.132820] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d866e5f5-3d83-4c03-a70d-9af52ea25867 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquiring lock "43db5f53-a719-4384-a8d2-63721dde0187-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.133077] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d866e5f5-3d83-4c03-a70d-9af52ea25867 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "43db5f53-a719-4384-a8d2-63721dde0187-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.133261] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d866e5f5-3d83-4c03-a70d-9af52ea25867 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "43db5f53-a719-4384-a8d2-63721dde0187-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.135431] env[62965]: INFO nova.compute.manager [None req-d866e5f5-3d83-4c03-a70d-9af52ea25867 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Terminating instance [ 1000.151861] env[62965]: DEBUG nova.objects.instance [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lazy-loading 'numa_topology' on Instance uuid 51babf43-b7b0-4731-a989-6977874fb801 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1000.200545] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a8075eef-2e02-46d9-83ab-33fe6536145e tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquiring lock "fe0559f0-d1ac-4af8-b554-2d31c38c2e73" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.200826] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a8075eef-2e02-46d9-83ab-33fe6536145e tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "fe0559f0-d1ac-4af8-b554-2d31c38c2e73" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.201070] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a8075eef-2e02-46d9-83ab-33fe6536145e tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquiring lock "fe0559f0-d1ac-4af8-b554-2d31c38c2e73-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.201287] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a8075eef-2e02-46d9-83ab-33fe6536145e tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "fe0559f0-d1ac-4af8-b554-2d31c38c2e73-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.201469] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a8075eef-2e02-46d9-83ab-33fe6536145e tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "fe0559f0-d1ac-4af8-b554-2d31c38c2e73-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.203940] env[62965]: INFO nova.compute.manager [None req-a8075eef-2e02-46d9-83ab-33fe6536145e tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Terminating instance [ 1000.249803] env[62965]: DEBUG nova.network.neutron [req-60f0d1db-05b5-4302-b625-ab15d8eec802 req-3011a1b5-5b89-4de7-b837-4ca1c1dfc3bd service nova] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Updated VIF entry in instance network info cache for port ee7ad38e-9936-45bf-8e20-b20e5f7958bd. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1000.250198] env[62965]: DEBUG nova.network.neutron [req-60f0d1db-05b5-4302-b625-ab15d8eec802 req-3011a1b5-5b89-4de7-b837-4ca1c1dfc3bd service nova] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Updating instance_info_cache with network_info: [{"id": "ee7ad38e-9936-45bf-8e20-b20e5f7958bd", "address": "fa:16:3e:93:d2:25", "network": {"id": "4649fed7-c2ca-43a9-a183-7c99423da726", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1978698173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fab418e2fe1420793517663574b43bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43ad01d2-c7dd-453c-a929-8ad76294d13c", "external-id": "nsx-vlan-transportzone-176", "segmentation_id": 176, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee7ad38e-99", "ovs_interfaceid": "ee7ad38e-9936-45bf-8e20-b20e5f7958bd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.318838] env[62965]: DEBUG oslo_vmware.api [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52c92024-6d43-c793-b5f1-f592f9f945bf, 'name': SearchDatastore_Task, 'duration_secs': 0.045018} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.319212] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.319469] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1000.319738] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.319920] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.320153] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1000.323241] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7925cfd2-d93b-4ba5-91af-3b155b00fd4e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.329110] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1000.329110] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1000.331820] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e983349-6216-4b96-ba1a-f8ddc819c80b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.337590] env[62965]: DEBUG oslo_vmware.api [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1000.337590] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]520950f8-b11e-d3cf-1503-9eac2f93655f" [ 1000.337590] env[62965]: _type = "Task" [ 1000.337590] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.345209] env[62965]: DEBUG oslo_vmware.api [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]520950f8-b11e-d3cf-1503-9eac2f93655f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.403893] env[62965]: DEBUG nova.network.neutron [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Updating instance_info_cache with network_info: [{"id": "170ea80e-f3c0-4647-af9d-34ffed538c33", "address": "fa:16:3e:23:92:be", "network": {"id": "a5246cc1-6001-4c20-9724-fd36a0437582", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-644702631-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bfa5cfdc7d5496482d87675f2a122dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap170ea80e-f3", "ovs_interfaceid": "170ea80e-f3c0-4647-af9d-34ffed538c33", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.638636] env[62965]: DEBUG nova.compute.manager [None req-d866e5f5-3d83-4c03-a70d-9af52ea25867 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1000.639144] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-d866e5f5-3d83-4c03-a70d-9af52ea25867 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1000.639780] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73fb0faf-cb3c-4abb-9892-6b6b73d871bf {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.647916] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-d866e5f5-3d83-4c03-a70d-9af52ea25867 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1000.648189] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dc3348dd-b114-4197-ae35-8088746e12e6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.654512] env[62965]: DEBUG nova.objects.base [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Object Instance<51babf43-b7b0-4731-a989-6977874fb801> lazy-loaded attributes: resources,numa_topology {{(pid=62965) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1000.656810] env[62965]: DEBUG oslo_vmware.api [None req-d866e5f5-3d83-4c03-a70d-9af52ea25867 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 1000.656810] env[62965]: value = "task-1390234" [ 1000.656810] env[62965]: _type = "Task" [ 1000.656810] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.665716] env[62965]: DEBUG oslo_vmware.api [None req-d866e5f5-3d83-4c03-a70d-9af52ea25867 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390234, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.708013] env[62965]: DEBUG nova.compute.manager [None req-a8075eef-2e02-46d9-83ab-33fe6536145e tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1000.708346] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-a8075eef-2e02-46d9-83ab-33fe6536145e tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1000.709294] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fea2d5d3-e021-4444-9338-a67168faaa78 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.719660] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8075eef-2e02-46d9-83ab-33fe6536145e tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1000.720143] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a66589a8-a3a7-4954-befa-126351ea84e7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.725685] env[62965]: DEBUG oslo_concurrency.lockutils [None req-500a830e-d2d3-465f-9a8b-456fd01cb73a tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquiring lock "refresh_cache-a0e9cc1a-bc95-4a44-99c5-aabf85a373f4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.725897] env[62965]: DEBUG oslo_concurrency.lockutils [None req-500a830e-d2d3-465f-9a8b-456fd01cb73a tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquired lock "refresh_cache-a0e9cc1a-bc95-4a44-99c5-aabf85a373f4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.726098] env[62965]: DEBUG nova.network.neutron [None req-500a830e-d2d3-465f-9a8b-456fd01cb73a tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1000.728784] env[62965]: DEBUG oslo_vmware.api [None req-a8075eef-2e02-46d9-83ab-33fe6536145e tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 1000.728784] env[62965]: value = "task-1390235" [ 1000.728784] env[62965]: _type = "Task" [ 1000.728784] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.739803] env[62965]: DEBUG oslo_vmware.api [None req-a8075eef-2e02-46d9-83ab-33fe6536145e tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390235, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.756220] env[62965]: DEBUG oslo_concurrency.lockutils [req-60f0d1db-05b5-4302-b625-ab15d8eec802 req-3011a1b5-5b89-4de7-b837-4ca1c1dfc3bd service nova] Releasing lock "refresh_cache-8efad776-6cbe-49ab-83dc-a410d8417505" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.756403] env[62965]: DEBUG nova.compute.manager [req-60f0d1db-05b5-4302-b625-ab15d8eec802 req-3011a1b5-5b89-4de7-b837-4ca1c1dfc3bd service nova] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Received event network-vif-unplugged-2664b4df-ad2e-4114-8837-6c5429ae94e3 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1000.756599] env[62965]: DEBUG oslo_concurrency.lockutils [req-60f0d1db-05b5-4302-b625-ab15d8eec802 req-3011a1b5-5b89-4de7-b837-4ca1c1dfc3bd service nova] Acquiring lock "51babf43-b7b0-4731-a989-6977874fb801-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.756803] env[62965]: DEBUG oslo_concurrency.lockutils [req-60f0d1db-05b5-4302-b625-ab15d8eec802 req-3011a1b5-5b89-4de7-b837-4ca1c1dfc3bd service nova] Lock "51babf43-b7b0-4731-a989-6977874fb801-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.756966] env[62965]: DEBUG oslo_concurrency.lockutils [req-60f0d1db-05b5-4302-b625-ab15d8eec802 req-3011a1b5-5b89-4de7-b837-4ca1c1dfc3bd service nova] Lock "51babf43-b7b0-4731-a989-6977874fb801-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.757150] env[62965]: DEBUG nova.compute.manager [req-60f0d1db-05b5-4302-b625-ab15d8eec802 req-3011a1b5-5b89-4de7-b837-4ca1c1dfc3bd service nova] [instance: 51babf43-b7b0-4731-a989-6977874fb801] No waiting events found dispatching network-vif-unplugged-2664b4df-ad2e-4114-8837-6c5429ae94e3 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1000.757323] env[62965]: WARNING nova.compute.manager [req-60f0d1db-05b5-4302-b625-ab15d8eec802 req-3011a1b5-5b89-4de7-b837-4ca1c1dfc3bd service nova] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Received unexpected event network-vif-unplugged-2664b4df-ad2e-4114-8837-6c5429ae94e3 for instance with vm_state shelved_offloaded and task_state None. [ 1000.757489] env[62965]: DEBUG nova.compute.manager [req-60f0d1db-05b5-4302-b625-ab15d8eec802 req-3011a1b5-5b89-4de7-b837-4ca1c1dfc3bd service nova] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Received event network-changed-2664b4df-ad2e-4114-8837-6c5429ae94e3 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1000.757645] env[62965]: DEBUG nova.compute.manager [req-60f0d1db-05b5-4302-b625-ab15d8eec802 req-3011a1b5-5b89-4de7-b837-4ca1c1dfc3bd service nova] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Refreshing instance network info cache due to event network-changed-2664b4df-ad2e-4114-8837-6c5429ae94e3. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 1000.757885] env[62965]: DEBUG oslo_concurrency.lockutils [req-60f0d1db-05b5-4302-b625-ab15d8eec802 req-3011a1b5-5b89-4de7-b837-4ca1c1dfc3bd service nova] Acquiring lock "refresh_cache-51babf43-b7b0-4731-a989-6977874fb801" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.758048] env[62965]: DEBUG oslo_concurrency.lockutils [req-60f0d1db-05b5-4302-b625-ab15d8eec802 req-3011a1b5-5b89-4de7-b837-4ca1c1dfc3bd service nova] Acquired lock "refresh_cache-51babf43-b7b0-4731-a989-6977874fb801" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.758228] env[62965]: DEBUG nova.network.neutron [req-60f0d1db-05b5-4302-b625-ab15d8eec802 req-3011a1b5-5b89-4de7-b837-4ca1c1dfc3bd service nova] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Refreshing network info cache for port 2664b4df-ad2e-4114-8837-6c5429ae94e3 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1000.818012] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-377412c9-7d87-49a5-b518-f41e96254f2c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.825699] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-924ec1da-64b2-4771-8394-2c1ea82e178a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.861897] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efbc8625-143c-4734-8649-8ad271b92d3d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.870642] env[62965]: DEBUG oslo_vmware.api [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]520950f8-b11e-d3cf-1503-9eac2f93655f, 'name': SearchDatastore_Task, 'duration_secs': 0.008137} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.873366] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bdc659fc-1f42-40d9-835d-6e18c671ed84 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.876467] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fa721fb-5400-48b4-99b7-7b671232b9d4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.883615] env[62965]: DEBUG oslo_vmware.api [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1000.883615] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]520c558e-5ed0-b25b-c60c-3be437614cfd" [ 1000.883615] env[62965]: _type = "Task" [ 1000.883615] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.892107] env[62965]: DEBUG nova.compute.provider_tree [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1000.903294] env[62965]: DEBUG oslo_vmware.api [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]520c558e-5ed0-b25b-c60c-3be437614cfd, 'name': SearchDatastore_Task, 'duration_secs': 0.010655} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.903294] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.903294] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] 8efad776-6cbe-49ab-83dc-a410d8417505/8efad776-6cbe-49ab-83dc-a410d8417505.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1000.903294] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-395a7f1a-e9e4-45e3-8a45-95cc0d8d106d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.906618] env[62965]: DEBUG oslo_concurrency.lockutils [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Releasing lock "refresh_cache-464a3a31-06bf-4662-a5b1-d185d9a64f0b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1000.914021] env[62965]: DEBUG oslo_vmware.api [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1000.914021] env[62965]: value = "task-1390236" [ 1000.914021] env[62965]: _type = "Task" [ 1000.914021] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.922358] env[62965]: DEBUG oslo_vmware.api [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390236, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.996905] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ca7b6073-f75d-485d-b9ca-fca9b05ed44b tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "51babf43-b7b0-4731-a989-6977874fb801" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.167688] env[62965]: DEBUG oslo_vmware.api [None req-d866e5f5-3d83-4c03-a70d-9af52ea25867 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390234, 'name': PowerOffVM_Task, 'duration_secs': 0.153957} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.167972] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-d866e5f5-3d83-4c03-a70d-9af52ea25867 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1001.168160] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-d866e5f5-3d83-4c03-a70d-9af52ea25867 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1001.168575] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d00e9f9c-62bf-48f6-b79c-a4890ca92e69 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.243255] env[62965]: DEBUG oslo_vmware.api [None req-a8075eef-2e02-46d9-83ab-33fe6536145e tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390235, 'name': PowerOffVM_Task, 'duration_secs': 0.152818} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.243537] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8075eef-2e02-46d9-83ab-33fe6536145e tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1001.243720] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-a8075eef-2e02-46d9-83ab-33fe6536145e tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1001.243968] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e34b7c29-0d08-405b-9382-dad298ee261d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.333716] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-d866e5f5-3d83-4c03-a70d-9af52ea25867 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1001.333930] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-d866e5f5-3d83-4c03-a70d-9af52ea25867 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Deleting contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1001.334136] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-d866e5f5-3d83-4c03-a70d-9af52ea25867 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Deleting the datastore file [datastore2] 43db5f53-a719-4384-a8d2-63721dde0187 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1001.334487] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-281cac8f-b072-4488-a757-50adcd176843 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.342719] env[62965]: DEBUG oslo_vmware.api [None req-d866e5f5-3d83-4c03-a70d-9af52ea25867 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 1001.342719] env[62965]: value = "task-1390239" [ 1001.342719] env[62965]: _type = "Task" [ 1001.342719] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.348807] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-a8075eef-2e02-46d9-83ab-33fe6536145e tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1001.349036] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-a8075eef-2e02-46d9-83ab-33fe6536145e tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Deleting contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1001.349222] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8075eef-2e02-46d9-83ab-33fe6536145e tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Deleting the datastore file [datastore2] fe0559f0-d1ac-4af8-b554-2d31c38c2e73 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1001.351440] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-313bf97b-3af4-45c7-a5ac-57c3a232b06b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.358044] env[62965]: DEBUG oslo_vmware.api [None req-d866e5f5-3d83-4c03-a70d-9af52ea25867 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390239, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.362587] env[62965]: DEBUG oslo_vmware.api [None req-a8075eef-2e02-46d9-83ab-33fe6536145e tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 1001.362587] env[62965]: value = "task-1390240" [ 1001.362587] env[62965]: _type = "Task" [ 1001.362587] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.369953] env[62965]: DEBUG oslo_vmware.api [None req-a8075eef-2e02-46d9-83ab-33fe6536145e tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390240, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.418491] env[62965]: ERROR nova.scheduler.client.report [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [req-f8741a11-a4c3-42b1-a18c-5dc2396fea95] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-f8741a11-a4c3-42b1-a18c-5dc2396fea95"}]} [ 1001.433091] env[62965]: DEBUG oslo_vmware.api [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390236, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.45191} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.433377] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] 8efad776-6cbe-49ab-83dc-a410d8417505/8efad776-6cbe-49ab-83dc-a410d8417505.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1001.433596] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1001.433894] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1006866c-0f7e-405a-a6bb-1d9f35ccd251 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.439392] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03ff17a2-5b36-4088-8ea9-3bd80963794a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.443793] env[62965]: DEBUG nova.scheduler.client.report [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Refreshing inventories for resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 1001.445703] env[62965]: DEBUG oslo_vmware.api [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1001.445703] env[62965]: value = "task-1390241" [ 1001.445703] env[62965]: _type = "Task" [ 1001.445703] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.467309] env[62965]: DEBUG nova.scheduler.client.report [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Updating ProviderTree inventory for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 1001.467530] env[62965]: DEBUG nova.compute.provider_tree [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1001.470305] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d26e8fcf-b6c6-4c45-aa54-9843b3445cf5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.478124] env[62965]: DEBUG oslo_vmware.api [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390241, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.482881] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Updating instance '464a3a31-06bf-4662-a5b1-d185d9a64f0b' progress to 83 {{(pid=62965) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1001.487597] env[62965]: DEBUG nova.scheduler.client.report [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Refreshing aggregate associations for resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8, aggregates: None {{(pid=62965) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 1001.507684] env[62965]: DEBUG nova.scheduler.client.report [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Refreshing trait associations for resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64 {{(pid=62965) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 1001.516523] env[62965]: DEBUG nova.network.neutron [None req-500a830e-d2d3-465f-9a8b-456fd01cb73a tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Updating instance_info_cache with network_info: [{"id": "6ffa156f-6a68-424e-b324-660892b9ad22", "address": "fa:16:3e:ed:d8:a1", "network": {"id": "e9a6ebdd-0479-45af-b947-5d35ae182c87", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-385052062-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89fd5275b2064288b44e79365f6f0271", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5ba07329-1d3e-4ba8-8774-d029262318c4", "external-id": "nsx-vlan-transportzone-534", "segmentation_id": 534, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ffa156f-6a", "ovs_interfaceid": "6ffa156f-6a68-424e-b324-660892b9ad22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.520396] env[62965]: DEBUG nova.network.neutron [req-60f0d1db-05b5-4302-b625-ab15d8eec802 req-3011a1b5-5b89-4de7-b837-4ca1c1dfc3bd service nova] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Updated VIF entry in instance network info cache for port 2664b4df-ad2e-4114-8837-6c5429ae94e3. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1001.520825] env[62965]: DEBUG nova.network.neutron [req-60f0d1db-05b5-4302-b625-ab15d8eec802 req-3011a1b5-5b89-4de7-b837-4ca1c1dfc3bd service nova] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Updating instance_info_cache with network_info: [{"id": "2664b4df-ad2e-4114-8837-6c5429ae94e3", "address": "fa:16:3e:52:04:a7", "network": {"id": "62052700-1781-4933-93c8-fa1aafdb0fd2", "bridge": null, "label": "tempest-DeleteServersTestJSON-1931798228-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "978666785b0743bb843de96585f96fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap2664b4df-ad", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.640034] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "0620c47c-2918-43dc-9e67-f32ca3dd20f1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.640871] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "0620c47c-2918-43dc-9e67-f32ca3dd20f1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.646997] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91b06ae2-c72b-4109-85b9-fc8c6b067bf4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.655521] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9198c396-3fdc-4150-875c-7e94c6953470 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.687603] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae242e23-2503-4ec0-b08a-a1789f9061e2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.694936] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9bcd949-496e-4f42-a54e-de397fc41331 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.708156] env[62965]: DEBUG nova.compute.provider_tree [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1001.852826] env[62965]: DEBUG oslo_vmware.api [None req-d866e5f5-3d83-4c03-a70d-9af52ea25867 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390239, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14892} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.853132] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-d866e5f5-3d83-4c03-a70d-9af52ea25867 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1001.853306] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-d866e5f5-3d83-4c03-a70d-9af52ea25867 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Deleted contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1001.853486] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-d866e5f5-3d83-4c03-a70d-9af52ea25867 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1001.853663] env[62965]: INFO nova.compute.manager [None req-d866e5f5-3d83-4c03-a70d-9af52ea25867 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Took 1.21 seconds to destroy the instance on the hypervisor. [ 1001.853900] env[62965]: DEBUG oslo.service.loopingcall [None req-d866e5f5-3d83-4c03-a70d-9af52ea25867 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1001.854106] env[62965]: DEBUG nova.compute.manager [-] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1001.854204] env[62965]: DEBUG nova.network.neutron [-] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1001.870694] env[62965]: DEBUG oslo_vmware.api [None req-a8075eef-2e02-46d9-83ab-33fe6536145e tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390240, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.159499} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.870914] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8075eef-2e02-46d9-83ab-33fe6536145e tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1001.871132] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-a8075eef-2e02-46d9-83ab-33fe6536145e tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Deleted contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1001.871286] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-a8075eef-2e02-46d9-83ab-33fe6536145e tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1001.871469] env[62965]: INFO nova.compute.manager [None req-a8075eef-2e02-46d9-83ab-33fe6536145e tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1001.871708] env[62965]: DEBUG oslo.service.loopingcall [None req-a8075eef-2e02-46d9-83ab-33fe6536145e tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1001.871894] env[62965]: DEBUG nova.compute.manager [-] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1001.871982] env[62965]: DEBUG nova.network.neutron [-] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1001.956233] env[62965]: DEBUG oslo_vmware.api [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390241, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073113} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.956517] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1001.957298] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9291a3e-1e8f-426a-a4aa-4ad3ae3c6532 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.978360] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] 8efad776-6cbe-49ab-83dc-a410d8417505/8efad776-6cbe-49ab-83dc-a410d8417505.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1001.978958] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9b370644-a02e-43a0-9df7-170bcc77ef32 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.994377] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1001.994650] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a245b436-1d6a-4aca-914f-2cffa0c6b57e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.001429] env[62965]: DEBUG oslo_vmware.api [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1002.001429] env[62965]: value = "task-1390242" [ 1002.001429] env[62965]: _type = "Task" [ 1002.001429] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.002624] env[62965]: DEBUG oslo_vmware.api [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 1002.002624] env[62965]: value = "task-1390243" [ 1002.002624] env[62965]: _type = "Task" [ 1002.002624] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.012841] env[62965]: DEBUG oslo_vmware.api [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390243, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.015869] env[62965]: DEBUG oslo_vmware.api [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390242, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.019460] env[62965]: DEBUG oslo_concurrency.lockutils [None req-500a830e-d2d3-465f-9a8b-456fd01cb73a tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Releasing lock "refresh_cache-a0e9cc1a-bc95-4a44-99c5-aabf85a373f4" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.020284] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-802ed628-a4a6-4507-81e8-ada1d8601fea {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.023083] env[62965]: DEBUG oslo_concurrency.lockutils [req-60f0d1db-05b5-4302-b625-ab15d8eec802 req-3011a1b5-5b89-4de7-b837-4ca1c1dfc3bd service nova] Releasing lock "refresh_cache-51babf43-b7b0-4731-a989-6977874fb801" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.027226] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-500a830e-d2d3-465f-9a8b-456fd01cb73a tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Resuming the VM {{(pid=62965) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1002.027476] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-246a5d0d-5bc2-4624-a2e8-404a21cb7b7c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.033305] env[62965]: DEBUG oslo_vmware.api [None req-500a830e-d2d3-465f-9a8b-456fd01cb73a tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 1002.033305] env[62965]: value = "task-1390244" [ 1002.033305] env[62965]: _type = "Task" [ 1002.033305] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.042961] env[62965]: DEBUG oslo_vmware.api [None req-500a830e-d2d3-465f-9a8b-456fd01cb73a tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390244, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.121322] env[62965]: DEBUG nova.compute.manager [req-f9782970-a310-46f4-bb90-1392e0eaef27 req-5bd84517-b200-429a-a55e-d482f74a613f service nova] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Received event network-vif-deleted-b70e2be1-f890-4164-9e52-1e4cceaa0be5 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1002.121463] env[62965]: INFO nova.compute.manager [req-f9782970-a310-46f4-bb90-1392e0eaef27 req-5bd84517-b200-429a-a55e-d482f74a613f service nova] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Neutron deleted interface b70e2be1-f890-4164-9e52-1e4cceaa0be5; detaching it from the instance and deleting it from the info cache [ 1002.121732] env[62965]: DEBUG nova.network.neutron [req-f9782970-a310-46f4-bb90-1392e0eaef27 req-5bd84517-b200-429a-a55e-d482f74a613f service nova] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.143011] env[62965]: DEBUG nova.compute.manager [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1002.161709] env[62965]: DEBUG nova.compute.manager [req-aff3c235-3e60-44da-b3c7-ed13a3c38140 req-556992d4-de22-41c5-a6df-273fb1b5dc11 service nova] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Received event network-vif-deleted-75d00c8d-3b3f-4a42-b9bb-219bbf284819 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1002.161975] env[62965]: INFO nova.compute.manager [req-aff3c235-3e60-44da-b3c7-ed13a3c38140 req-556992d4-de22-41c5-a6df-273fb1b5dc11 service nova] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Neutron deleted interface 75d00c8d-3b3f-4a42-b9bb-219bbf284819; detaching it from the instance and deleting it from the info cache [ 1002.162247] env[62965]: DEBUG nova.network.neutron [req-aff3c235-3e60-44da-b3c7-ed13a3c38140 req-556992d4-de22-41c5-a6df-273fb1b5dc11 service nova] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.211434] env[62965]: DEBUG nova.scheduler.client.report [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1002.519074] env[62965]: DEBUG oslo_vmware.api [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390243, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.522198] env[62965]: DEBUG oslo_vmware.api [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390242, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.546649] env[62965]: DEBUG oslo_vmware.api [None req-500a830e-d2d3-465f-9a8b-456fd01cb73a tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390244, 'name': PowerOnVM_Task} progress is 93%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.590531] env[62965]: DEBUG nova.network.neutron [-] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.627800] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3538cc7e-ac8c-4fec-be7c-036ae35b887b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.630079] env[62965]: DEBUG nova.network.neutron [-] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.641073] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6153181d-fafa-47dc-a6d4-afa62949f3c0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.664944] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c0c7ff7d-17f3-42ae-b51a-1e1433d8966c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.676160] env[62965]: DEBUG nova.compute.manager [req-f9782970-a310-46f4-bb90-1392e0eaef27 req-5bd84517-b200-429a-a55e-d482f74a613f service nova] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Detach interface failed, port_id=b70e2be1-f890-4164-9e52-1e4cceaa0be5, reason: Instance 43db5f53-a719-4384-a8d2-63721dde0187 could not be found. {{(pid=62965) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11294}} [ 1002.677458] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.687315] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61951b64-ee05-419a-af06-8156e7145a3a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.721127] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.072s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.723760] env[62965]: DEBUG nova.compute.manager [req-aff3c235-3e60-44da-b3c7-ed13a3c38140 req-556992d4-de22-41c5-a6df-273fb1b5dc11 service nova] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Detach interface failed, port_id=75d00c8d-3b3f-4a42-b9bb-219bbf284819, reason: Instance fe0559f0-d1ac-4af8-b554-2d31c38c2e73 could not be found. {{(pid=62965) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11294}} [ 1002.724502] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.047s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.725969] env[62965]: INFO nova.compute.claims [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1003.015366] env[62965]: DEBUG oslo_vmware.api [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390242, 'name': ReconfigVM_Task, 'duration_secs': 0.534704} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.018306] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Reconfigured VM instance instance-0000005f to attach disk [datastore1] 8efad776-6cbe-49ab-83dc-a410d8417505/8efad776-6cbe-49ab-83dc-a410d8417505.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1003.018984] env[62965]: DEBUG oslo_vmware.api [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390243, 'name': PowerOnVM_Task, 'duration_secs': 0.529552} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.019218] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b53ac168-59fd-4ca7-8ce8-f479298f0622 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.020641] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1003.020846] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-bee0f087-77b6-4d20-ab0b-43bdc4400d4f tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Updating instance '464a3a31-06bf-4662-a5b1-d185d9a64f0b' progress to 100 {{(pid=62965) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1003.028439] env[62965]: DEBUG oslo_vmware.api [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1003.028439] env[62965]: value = "task-1390246" [ 1003.028439] env[62965]: _type = "Task" [ 1003.028439] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.036292] env[62965]: DEBUG oslo_vmware.api [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390246, 'name': Rename_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.043219] env[62965]: DEBUG oslo_vmware.api [None req-500a830e-d2d3-465f-9a8b-456fd01cb73a tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390244, 'name': PowerOnVM_Task, 'duration_secs': 0.616014} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.043457] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-500a830e-d2d3-465f-9a8b-456fd01cb73a tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Resumed the VM {{(pid=62965) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1003.043678] env[62965]: DEBUG nova.compute.manager [None req-500a830e-d2d3-465f-9a8b-456fd01cb73a tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1003.044395] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa0a5e55-2ed0-4a21-b314-4cc34be3a02a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.093370] env[62965]: INFO nova.compute.manager [-] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Took 1.24 seconds to deallocate network for instance. [ 1003.133062] env[62965]: INFO nova.compute.manager [-] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Took 1.26 seconds to deallocate network for instance. [ 1003.231818] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d16d3311-ead9-4778-9ee8-4f0bfe50ba95 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "51babf43-b7b0-4731-a989-6977874fb801" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 21.875s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.232877] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ca7b6073-f75d-485d-b9ca-fca9b05ed44b tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "51babf43-b7b0-4731-a989-6977874fb801" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 2.236s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.233252] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ca7b6073-f75d-485d-b9ca-fca9b05ed44b tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "51babf43-b7b0-4731-a989-6977874fb801-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.233352] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ca7b6073-f75d-485d-b9ca-fca9b05ed44b tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "51babf43-b7b0-4731-a989-6977874fb801-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.233501] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ca7b6073-f75d-485d-b9ca-fca9b05ed44b tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "51babf43-b7b0-4731-a989-6977874fb801-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.235294] env[62965]: INFO nova.compute.manager [None req-ca7b6073-f75d-485d-b9ca-fca9b05ed44b tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Terminating instance [ 1003.542886] env[62965]: DEBUG oslo_vmware.api [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390246, 'name': Rename_Task, 'duration_secs': 0.231219} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.543283] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1003.543629] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b479af97-cb95-4893-913d-bbd6c65a43f1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.552842] env[62965]: DEBUG oslo_vmware.api [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1003.552842] env[62965]: value = "task-1390247" [ 1003.552842] env[62965]: _type = "Task" [ 1003.552842] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.570188] env[62965]: DEBUG oslo_vmware.api [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390247, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.599562] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d866e5f5-3d83-4c03-a70d-9af52ea25867 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.638817] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a8075eef-2e02-46d9-83ab-33fe6536145e tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.738918] env[62965]: DEBUG nova.compute.manager [None req-ca7b6073-f75d-485d-b9ca-fca9b05ed44b tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1003.739354] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ca7b6073-f75d-485d-b9ca-fca9b05ed44b tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1003.739408] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-66f9d40d-f1eb-4fea-9bae-4a3a84b0b8ad {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.749010] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0da1f166-aead-439a-bff5-319f80ea9900 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.785798] env[62965]: WARNING nova.virt.vmwareapi.vmops [None req-ca7b6073-f75d-485d-b9ca-fca9b05ed44b tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 51babf43-b7b0-4731-a989-6977874fb801 could not be found. [ 1003.786096] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ca7b6073-f75d-485d-b9ca-fca9b05ed44b tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1003.786291] env[62965]: INFO nova.compute.manager [None req-ca7b6073-f75d-485d-b9ca-fca9b05ed44b tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Took 0.05 seconds to destroy the instance on the hypervisor. [ 1003.786548] env[62965]: DEBUG oslo.service.loopingcall [None req-ca7b6073-f75d-485d-b9ca-fca9b05ed44b tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1003.789180] env[62965]: DEBUG nova.compute.manager [-] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1003.789283] env[62965]: DEBUG nova.network.neutron [-] [instance: 51babf43-b7b0-4731-a989-6977874fb801] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1003.911784] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fe37e85-7ba8-483e-958e-e36193943b62 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.920100] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfff6979-ae55-42f5-8a58-66791fbf4fa8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.949336] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5034dc8-be33-4907-9077-b8d6fa172294 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.956971] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b98f156-2eff-4462-96a0-a2ccb2388ac7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.970733] env[62965]: DEBUG nova.compute.provider_tree [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1004.062353] env[62965]: DEBUG oslo_vmware.api [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390247, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.443603] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-06d11f1e-7c69-424b-9cfb-f1be5384bcba tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Volume attach. Driver type: vmdk {{(pid=62965) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1004.443858] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-06d11f1e-7c69-424b-9cfb-f1be5384bcba tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-295045', 'volume_id': 'bbfeace6-81aa-40a3-8563-95c2c9a1c7ef', 'name': 'volume-bbfeace6-81aa-40a3-8563-95c2c9a1c7ef', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a63b301d-8d52-4fec-951f-a15cda9449ac', 'attached_at': '', 'detached_at': '', 'volume_id': 'bbfeace6-81aa-40a3-8563-95c2c9a1c7ef', 'serial': 'bbfeace6-81aa-40a3-8563-95c2c9a1c7ef'} {{(pid=62965) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1004.444760] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37758bdf-a568-4663-a8aa-f05267d967ea {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.461595] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b567be3a-dbf1-4048-add2-cebc66e39b3a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.479324] env[62965]: DEBUG nova.scheduler.client.report [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1004.489869] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-06d11f1e-7c69-424b-9cfb-f1be5384bcba tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] volume-bbfeace6-81aa-40a3-8563-95c2c9a1c7ef/volume-bbfeace6-81aa-40a3-8563-95c2c9a1c7ef.vmdk or device None with type thin {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1004.490437] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7cfdd827-8db4-4863-9405-7afa52f3b99f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.509022] env[62965]: DEBUG oslo_vmware.api [None req-06d11f1e-7c69-424b-9cfb-f1be5384bcba tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Waiting for the task: (returnval){ [ 1004.509022] env[62965]: value = "task-1390248" [ 1004.509022] env[62965]: _type = "Task" [ 1004.509022] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.517229] env[62965]: DEBUG oslo_vmware.api [None req-06d11f1e-7c69-424b-9cfb-f1be5384bcba tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390248, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.566817] env[62965]: DEBUG oslo_vmware.api [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390247, 'name': PowerOnVM_Task} progress is 81%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.637845] env[62965]: DEBUG nova.network.neutron [-] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.991566] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.267s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.992115] env[62965]: DEBUG nova.compute.manager [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1004.994708] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d866e5f5-3d83-4c03-a70d-9af52ea25867 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.395s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.994918] env[62965]: DEBUG nova.objects.instance [None req-d866e5f5-3d83-4c03-a70d-9af52ea25867 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lazy-loading 'resources' on Instance uuid 43db5f53-a719-4384-a8d2-63721dde0187 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1005.019849] env[62965]: DEBUG oslo_vmware.api [None req-06d11f1e-7c69-424b-9cfb-f1be5384bcba tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390248, 'name': ReconfigVM_Task, 'duration_secs': 0.393853} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.021028] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-06d11f1e-7c69-424b-9cfb-f1be5384bcba tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Reconfigured VM instance instance-00000058 to attach disk [datastore1] volume-bbfeace6-81aa-40a3-8563-95c2c9a1c7ef/volume-bbfeace6-81aa-40a3-8563-95c2c9a1c7ef.vmdk or device None with type thin {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1005.025429] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f81b0a61-0052-418f-bb73-8eb6b4860e03 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.041534] env[62965]: DEBUG oslo_vmware.api [None req-06d11f1e-7c69-424b-9cfb-f1be5384bcba tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Waiting for the task: (returnval){ [ 1005.041534] env[62965]: value = "task-1390249" [ 1005.041534] env[62965]: _type = "Task" [ 1005.041534] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.049749] env[62965]: DEBUG oslo_vmware.api [None req-06d11f1e-7c69-424b-9cfb-f1be5384bcba tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390249, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.061901] env[62965]: DEBUG oslo_vmware.api [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390247, 'name': PowerOnVM_Task, 'duration_secs': 1.167328} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.062177] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1005.062407] env[62965]: INFO nova.compute.manager [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Took 7.83 seconds to spawn the instance on the hypervisor. [ 1005.062713] env[62965]: DEBUG nova.compute.manager [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1005.063374] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79d1aff4-6b19-4178-abad-82bf3f4e8f78 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.140657] env[62965]: INFO nova.compute.manager [-] [instance: 51babf43-b7b0-4731-a989-6977874fb801] Took 1.35 seconds to deallocate network for instance. [ 1005.266960] env[62965]: DEBUG oslo_concurrency.lockutils [None req-01458532-419b-4fa3-bf4c-ef681d3f6607 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "464a3a31-06bf-4662-a5b1-d185d9a64f0b" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.267269] env[62965]: DEBUG oslo_concurrency.lockutils [None req-01458532-419b-4fa3-bf4c-ef681d3f6607 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "464a3a31-06bf-4662-a5b1-d185d9a64f0b" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.267460] env[62965]: DEBUG nova.compute.manager [None req-01458532-419b-4fa3-bf4c-ef681d3f6607 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Going to confirm migration 2 {{(pid=62965) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5112}} [ 1005.498250] env[62965]: DEBUG nova.compute.utils [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1005.500564] env[62965]: DEBUG nova.compute.manager [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1005.500564] env[62965]: DEBUG nova.network.neutron [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1005.555129] env[62965]: DEBUG oslo_vmware.api [None req-06d11f1e-7c69-424b-9cfb-f1be5384bcba tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390249, 'name': ReconfigVM_Task, 'duration_secs': 0.151374} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.555627] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-06d11f1e-7c69-424b-9cfb-f1be5384bcba tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-295045', 'volume_id': 'bbfeace6-81aa-40a3-8563-95c2c9a1c7ef', 'name': 'volume-bbfeace6-81aa-40a3-8563-95c2c9a1c7ef', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a63b301d-8d52-4fec-951f-a15cda9449ac', 'attached_at': '', 'detached_at': '', 'volume_id': 'bbfeace6-81aa-40a3-8563-95c2c9a1c7ef', 'serial': 'bbfeace6-81aa-40a3-8563-95c2c9a1c7ef'} {{(pid=62965) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1005.584200] env[62965]: INFO nova.compute.manager [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Took 13.71 seconds to build instance. [ 1005.591309] env[62965]: DEBUG nova.policy [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '97ee4fdb7cdb45829c7fe694e883da8c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1ed71548c085499981fa1b102d660368', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 1005.684580] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1ad502a-a8cb-408b-af6f-cb1d7f48b3be {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.692204] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31f65778-de58-4813-a00f-fcc5dea2e5b8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.738725] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf2d00b8-bd28-44b8-a47a-abec2c980a56 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.746310] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33e48fa8-063c-40df-afa8-0fef03542552 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.759486] env[62965]: DEBUG nova.compute.provider_tree [None req-d866e5f5-3d83-4c03-a70d-9af52ea25867 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1005.827817] env[62965]: DEBUG oslo_concurrency.lockutils [None req-01458532-419b-4fa3-bf4c-ef681d3f6607 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "refresh_cache-464a3a31-06bf-4662-a5b1-d185d9a64f0b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.827997] env[62965]: DEBUG oslo_concurrency.lockutils [None req-01458532-419b-4fa3-bf4c-ef681d3f6607 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquired lock "refresh_cache-464a3a31-06bf-4662-a5b1-d185d9a64f0b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.828184] env[62965]: DEBUG nova.network.neutron [None req-01458532-419b-4fa3-bf4c-ef681d3f6607 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1005.828366] env[62965]: DEBUG nova.objects.instance [None req-01458532-419b-4fa3-bf4c-ef681d3f6607 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lazy-loading 'info_cache' on Instance uuid 464a3a31-06bf-4662-a5b1-d185d9a64f0b {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1006.002868] env[62965]: DEBUG nova.compute.manager [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1006.034722] env[62965]: INFO nova.compute.manager [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Rebuilding instance [ 1006.086644] env[62965]: DEBUG nova.compute.manager [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1006.087118] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5aa83e9c-f6a0-444f-aead-6f6a65f31a21 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "8efad776-6cbe-49ab-83dc-a410d8417505" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.243s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.087960] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab0a8316-60de-4d03-a0b5-5303fe29ef1a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.144738] env[62965]: DEBUG nova.network.neutron [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Successfully created port: b178e278-58df-4607-ae3b-1d6c07cd22e7 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1006.177654] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ca7b6073-f75d-485d-b9ca-fca9b05ed44b tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "51babf43-b7b0-4731-a989-6977874fb801" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.945s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.263111] env[62965]: DEBUG nova.scheduler.client.report [None req-d866e5f5-3d83-4c03-a70d-9af52ea25867 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1006.600736] env[62965]: DEBUG nova.objects.instance [None req-06d11f1e-7c69-424b-9cfb-f1be5384bcba tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lazy-loading 'flavor' on Instance uuid a63b301d-8d52-4fec-951f-a15cda9449ac {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1006.769667] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d866e5f5-3d83-4c03-a70d-9af52ea25867 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.775s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.772080] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a8075eef-2e02-46d9-83ab-33fe6536145e tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.133s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.772375] env[62965]: DEBUG nova.objects.instance [None req-a8075eef-2e02-46d9-83ab-33fe6536145e tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lazy-loading 'resources' on Instance uuid fe0559f0-d1ac-4af8-b554-2d31c38c2e73 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1006.786726] env[62965]: INFO nova.scheduler.client.report [None req-d866e5f5-3d83-4c03-a70d-9af52ea25867 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Deleted allocations for instance 43db5f53-a719-4384-a8d2-63721dde0187 [ 1007.011739] env[62965]: DEBUG nova.compute.manager [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1007.034238] env[62965]: DEBUG nova.network.neutron [None req-01458532-419b-4fa3-bf4c-ef681d3f6607 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Updating instance_info_cache with network_info: [{"id": "170ea80e-f3c0-4647-af9d-34ffed538c33", "address": "fa:16:3e:23:92:be", "network": {"id": "a5246cc1-6001-4c20-9724-fd36a0437582", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-644702631-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bfa5cfdc7d5496482d87675f2a122dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap170ea80e-f3", "ovs_interfaceid": "170ea80e-f3c0-4647-af9d-34ffed538c33", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.039839] env[62965]: DEBUG nova.virt.hardware [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1007.040070] env[62965]: DEBUG nova.virt.hardware [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1007.040237] env[62965]: DEBUG nova.virt.hardware [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1007.040421] env[62965]: DEBUG nova.virt.hardware [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1007.040592] env[62965]: DEBUG nova.virt.hardware [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1007.040756] env[62965]: DEBUG nova.virt.hardware [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1007.040963] env[62965]: DEBUG nova.virt.hardware [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1007.041143] env[62965]: DEBUG nova.virt.hardware [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1007.041315] env[62965]: DEBUG nova.virt.hardware [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1007.041499] env[62965]: DEBUG nova.virt.hardware [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1007.041711] env[62965]: DEBUG nova.virt.hardware [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1007.042647] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-829bdd52-a69a-4801-a8db-f62958211580 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.050831] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4cf7311-6819-47c4-a7fa-46e170d4057e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.103959] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1007.105471] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ee9b5204-d740-4419-871f-01bb1d049d10 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.107469] env[62965]: DEBUG oslo_concurrency.lockutils [None req-06d11f1e-7c69-424b-9cfb-f1be5384bcba tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "a63b301d-8d52-4fec-951f-a15cda9449ac" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.264s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.114311] env[62965]: DEBUG oslo_vmware.api [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1007.114311] env[62965]: value = "task-1390250" [ 1007.114311] env[62965]: _type = "Task" [ 1007.114311] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.124035] env[62965]: DEBUG oslo_vmware.api [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390250, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.277527] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "8ae1fde5-e94e-43ed-816a-9e22633dfdd1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.277757] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "8ae1fde5-e94e-43ed-816a-9e22633dfdd1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.294277] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d866e5f5-3d83-4c03-a70d-9af52ea25867 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "43db5f53-a719-4384-a8d2-63721dde0187" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.162s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.428748] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83231340-7e19-442c-b37c-ea46433c18e3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.436844] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6e25880-955e-478e-9eb4-35982a3d37f1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.469182] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2b3fc472-7a55-4324-9044-8e770a80c956 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquiring lock "a63b301d-8d52-4fec-951f-a15cda9449ac" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.469444] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2b3fc472-7a55-4324-9044-8e770a80c956 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "a63b301d-8d52-4fec-951f-a15cda9449ac" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.471563] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcfac153-1233-4b3f-b603-234acbdcad53 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.479945] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24f8f1df-c312-4230-9acf-a1a82c4ab070 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.494286] env[62965]: DEBUG nova.compute.provider_tree [None req-a8075eef-2e02-46d9-83ab-33fe6536145e tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1007.536538] env[62965]: DEBUG oslo_concurrency.lockutils [None req-01458532-419b-4fa3-bf4c-ef681d3f6607 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Releasing lock "refresh_cache-464a3a31-06bf-4662-a5b1-d185d9a64f0b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.536737] env[62965]: DEBUG nova.objects.instance [None req-01458532-419b-4fa3-bf4c-ef681d3f6607 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lazy-loading 'migration_context' on Instance uuid 464a3a31-06bf-4662-a5b1-d185d9a64f0b {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1007.560338] env[62965]: DEBUG nova.compute.manager [req-16a07859-5307-415e-910f-5556b217c286 req-34d935fd-92c6-4f4d-b375-3af61f94ff01 service nova] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Received event network-vif-plugged-b178e278-58df-4607-ae3b-1d6c07cd22e7 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1007.560605] env[62965]: DEBUG oslo_concurrency.lockutils [req-16a07859-5307-415e-910f-5556b217c286 req-34d935fd-92c6-4f4d-b375-3af61f94ff01 service nova] Acquiring lock "0620c47c-2918-43dc-9e67-f32ca3dd20f1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.560829] env[62965]: DEBUG oslo_concurrency.lockutils [req-16a07859-5307-415e-910f-5556b217c286 req-34d935fd-92c6-4f4d-b375-3af61f94ff01 service nova] Lock "0620c47c-2918-43dc-9e67-f32ca3dd20f1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.560998] env[62965]: DEBUG oslo_concurrency.lockutils [req-16a07859-5307-415e-910f-5556b217c286 req-34d935fd-92c6-4f4d-b375-3af61f94ff01 service nova] Lock "0620c47c-2918-43dc-9e67-f32ca3dd20f1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.561185] env[62965]: DEBUG nova.compute.manager [req-16a07859-5307-415e-910f-5556b217c286 req-34d935fd-92c6-4f4d-b375-3af61f94ff01 service nova] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] No waiting events found dispatching network-vif-plugged-b178e278-58df-4607-ae3b-1d6c07cd22e7 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1007.561354] env[62965]: WARNING nova.compute.manager [req-16a07859-5307-415e-910f-5556b217c286 req-34d935fd-92c6-4f4d-b375-3af61f94ff01 service nova] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Received unexpected event network-vif-plugged-b178e278-58df-4607-ae3b-1d6c07cd22e7 for instance with vm_state building and task_state spawning. [ 1007.625019] env[62965]: DEBUG oslo_vmware.api [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390250, 'name': PowerOffVM_Task, 'duration_secs': 0.17892} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.625379] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1007.625620] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1007.627223] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a7bafec-972c-4e8b-8d7c-a3a224e4dbb0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.634860] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1007.635138] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c60a06ce-eca4-461b-b094-386a10269101 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.705267] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1007.705507] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Deleting contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1007.705690] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Deleting the datastore file [datastore1] 8efad776-6cbe-49ab-83dc-a410d8417505 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1007.705963] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0c892068-7c3d-4bd1-ad74-54964d4cb248 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.712723] env[62965]: DEBUG oslo_vmware.api [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1007.712723] env[62965]: value = "task-1390252" [ 1007.712723] env[62965]: _type = "Task" [ 1007.712723] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.721218] env[62965]: DEBUG oslo_vmware.api [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390252, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.780971] env[62965]: DEBUG nova.compute.manager [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1007.976499] env[62965]: INFO nova.compute.manager [None req-2b3fc472-7a55-4324-9044-8e770a80c956 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Detaching volume bbfeace6-81aa-40a3-8563-95c2c9a1c7ef [ 1007.997761] env[62965]: DEBUG nova.scheduler.client.report [None req-a8075eef-2e02-46d9-83ab-33fe6536145e tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1008.015617] env[62965]: INFO nova.virt.block_device [None req-2b3fc472-7a55-4324-9044-8e770a80c956 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Attempting to driver detach volume bbfeace6-81aa-40a3-8563-95c2c9a1c7ef from mountpoint /dev/sdb [ 1008.015934] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b3fc472-7a55-4324-9044-8e770a80c956 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Volume detach. Driver type: vmdk {{(pid=62965) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1008.016178] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b3fc472-7a55-4324-9044-8e770a80c956 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-295045', 'volume_id': 'bbfeace6-81aa-40a3-8563-95c2c9a1c7ef', 'name': 'volume-bbfeace6-81aa-40a3-8563-95c2c9a1c7ef', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a63b301d-8d52-4fec-951f-a15cda9449ac', 'attached_at': '', 'detached_at': '', 'volume_id': 'bbfeace6-81aa-40a3-8563-95c2c9a1c7ef', 'serial': 'bbfeace6-81aa-40a3-8563-95c2c9a1c7ef'} {{(pid=62965) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1008.017135] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7258aafc-000e-47c6-a0f9-fbafea08037a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.040076] env[62965]: DEBUG nova.objects.base [None req-01458532-419b-4fa3-bf4c-ef681d3f6607 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Object Instance<464a3a31-06bf-4662-a5b1-d185d9a64f0b> lazy-loaded attributes: info_cache,migration_context {{(pid=62965) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1008.040975] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9199a7e1-c260-4db5-8259-b0ce5594186f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.043965] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5abeb4c9-67db-4881-8815-7b2044cf43b3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.065846] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec0623c6-f800-483b-9643-91aa2813bf67 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.068407] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0de4034-6897-4a92-94da-0e0076889ef0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.073521] env[62965]: DEBUG oslo_vmware.api [None req-01458532-419b-4fa3-bf4c-ef681d3f6607 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 1008.073521] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]527a4045-b941-b2f5-adc9-36f90d704444" [ 1008.073521] env[62965]: _type = "Task" [ 1008.073521] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.092684] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f235781-a849-43c8-b23a-8f4b9d30a47a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.097949] env[62965]: DEBUG oslo_vmware.api [None req-01458532-419b-4fa3-bf4c-ef681d3f6607 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]527a4045-b941-b2f5-adc9-36f90d704444, 'name': SearchDatastore_Task, 'duration_secs': 0.006409} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.098535] env[62965]: DEBUG oslo_concurrency.lockutils [None req-01458532-419b-4fa3-bf4c-ef681d3f6607 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.109612] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b3fc472-7a55-4324-9044-8e770a80c956 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] The volume has not been displaced from its original location: [datastore1] volume-bbfeace6-81aa-40a3-8563-95c2c9a1c7ef/volume-bbfeace6-81aa-40a3-8563-95c2c9a1c7ef.vmdk. No consolidation needed. {{(pid=62965) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1008.114836] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b3fc472-7a55-4324-9044-8e770a80c956 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Reconfiguring VM instance instance-00000058 to detach disk 2001 {{(pid=62965) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1008.115130] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6a4332aa-6c44-4184-a23f-2b1bfbb791a7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.129547] env[62965]: DEBUG nova.network.neutron [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Successfully updated port: b178e278-58df-4607-ae3b-1d6c07cd22e7 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1008.135345] env[62965]: DEBUG oslo_vmware.api [None req-2b3fc472-7a55-4324-9044-8e770a80c956 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Waiting for the task: (returnval){ [ 1008.135345] env[62965]: value = "task-1390253" [ 1008.135345] env[62965]: _type = "Task" [ 1008.135345] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.144603] env[62965]: DEBUG oslo_vmware.api [None req-2b3fc472-7a55-4324-9044-8e770a80c956 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390253, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.147070] env[62965]: DEBUG nova.compute.manager [req-fbea45f8-c638-4fd9-b825-7f8231818b7d req-cf85fe2d-28f3-4770-a867-f6ba8b70b011 service nova] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Received event network-changed-b178e278-58df-4607-ae3b-1d6c07cd22e7 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1008.147269] env[62965]: DEBUG nova.compute.manager [req-fbea45f8-c638-4fd9-b825-7f8231818b7d req-cf85fe2d-28f3-4770-a867-f6ba8b70b011 service nova] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Refreshing instance network info cache due to event network-changed-b178e278-58df-4607-ae3b-1d6c07cd22e7. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 1008.147488] env[62965]: DEBUG oslo_concurrency.lockutils [req-fbea45f8-c638-4fd9-b825-7f8231818b7d req-cf85fe2d-28f3-4770-a867-f6ba8b70b011 service nova] Acquiring lock "refresh_cache-0620c47c-2918-43dc-9e67-f32ca3dd20f1" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.147699] env[62965]: DEBUG oslo_concurrency.lockutils [req-fbea45f8-c638-4fd9-b825-7f8231818b7d req-cf85fe2d-28f3-4770-a867-f6ba8b70b011 service nova] Acquired lock "refresh_cache-0620c47c-2918-43dc-9e67-f32ca3dd20f1" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.147777] env[62965]: DEBUG nova.network.neutron [req-fbea45f8-c638-4fd9-b825-7f8231818b7d req-cf85fe2d-28f3-4770-a867-f6ba8b70b011 service nova] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Refreshing network info cache for port b178e278-58df-4607-ae3b-1d6c07cd22e7 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1008.224328] env[62965]: DEBUG oslo_vmware.api [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390252, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.142496} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.224543] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1008.224736] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Deleted contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1008.224917] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1008.307850] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.502909] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a8075eef-2e02-46d9-83ab-33fe6536145e tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.731s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.506344] env[62965]: DEBUG oslo_concurrency.lockutils [None req-01458532-419b-4fa3-bf4c-ef681d3f6607 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.407s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.522422] env[62965]: INFO nova.scheduler.client.report [None req-a8075eef-2e02-46d9-83ab-33fe6536145e tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Deleted allocations for instance fe0559f0-d1ac-4af8-b554-2d31c38c2e73 [ 1008.630390] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "refresh_cache-0620c47c-2918-43dc-9e67-f32ca3dd20f1" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.644961] env[62965]: DEBUG oslo_vmware.api [None req-2b3fc472-7a55-4324-9044-8e770a80c956 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390253, 'name': ReconfigVM_Task, 'duration_secs': 0.240154} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.645301] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b3fc472-7a55-4324-9044-8e770a80c956 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Reconfigured VM instance instance-00000058 to detach disk 2001 {{(pid=62965) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1008.652054] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-90ea1bed-bf7c-4497-8b38-7c0c70404933 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.667955] env[62965]: DEBUG oslo_vmware.api [None req-2b3fc472-7a55-4324-9044-8e770a80c956 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Waiting for the task: (returnval){ [ 1008.667955] env[62965]: value = "task-1390254" [ 1008.667955] env[62965]: _type = "Task" [ 1008.667955] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.675814] env[62965]: DEBUG oslo_vmware.api [None req-2b3fc472-7a55-4324-9044-8e770a80c956 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390254, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.695763] env[62965]: DEBUG nova.network.neutron [req-fbea45f8-c638-4fd9-b825-7f8231818b7d req-cf85fe2d-28f3-4770-a867-f6ba8b70b011 service nova] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1008.794256] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7e818b1d-898d-46dc-9cff-8175b4ed1ce6 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquiring lock "a0e9cc1a-bc95-4a44-99c5-aabf85a373f4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.794637] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7e818b1d-898d-46dc-9cff-8175b4ed1ce6 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lock "a0e9cc1a-bc95-4a44-99c5-aabf85a373f4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.794870] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7e818b1d-898d-46dc-9cff-8175b4ed1ce6 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquiring lock "a0e9cc1a-bc95-4a44-99c5-aabf85a373f4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.795087] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7e818b1d-898d-46dc-9cff-8175b4ed1ce6 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lock "a0e9cc1a-bc95-4a44-99c5-aabf85a373f4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.795248] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7e818b1d-898d-46dc-9cff-8175b4ed1ce6 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lock "a0e9cc1a-bc95-4a44-99c5-aabf85a373f4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.797407] env[62965]: INFO nova.compute.manager [None req-7e818b1d-898d-46dc-9cff-8175b4ed1ce6 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Terminating instance [ 1008.829805] env[62965]: DEBUG nova.network.neutron [req-fbea45f8-c638-4fd9-b825-7f8231818b7d req-cf85fe2d-28f3-4770-a867-f6ba8b70b011 service nova] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.029846] env[62965]: DEBUG oslo_concurrency.lockutils [None req-a8075eef-2e02-46d9-83ab-33fe6536145e tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "fe0559f0-d1ac-4af8-b554-2d31c38c2e73" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.829s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.161037] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce9e53d6-999f-4e5c-a50d-ebada574bd36 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.169043] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39033c2f-fd87-4c82-accd-cf33ea8bf0da {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.179383] env[62965]: DEBUG oslo_vmware.api [None req-2b3fc472-7a55-4324-9044-8e770a80c956 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390254, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.204066] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05647783-7c17-45d8-9ab7-1309c20c50b2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.210952] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04db06c0-9a37-456d-80e8-85e413227464 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.223539] env[62965]: DEBUG nova.compute.provider_tree [None req-01458532-419b-4fa3-bf4c-ef681d3f6607 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1009.256593] env[62965]: DEBUG nova.virt.hardware [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1009.256829] env[62965]: DEBUG nova.virt.hardware [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1009.256987] env[62965]: DEBUG nova.virt.hardware [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1009.257185] env[62965]: DEBUG nova.virt.hardware [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1009.257333] env[62965]: DEBUG nova.virt.hardware [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1009.257481] env[62965]: DEBUG nova.virt.hardware [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1009.257686] env[62965]: DEBUG nova.virt.hardware [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1009.257850] env[62965]: DEBUG nova.virt.hardware [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1009.258030] env[62965]: DEBUG nova.virt.hardware [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1009.258198] env[62965]: DEBUG nova.virt.hardware [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1009.258372] env[62965]: DEBUG nova.virt.hardware [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1009.259169] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78841a4e-daa3-41cb-9599-ff1628d673f6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.267131] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7159bea4-89ce-47e1-a1df-aac235f798bd {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.279732] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:93:d2:25', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '43ad01d2-c7dd-453c-a929-8ad76294d13c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ee7ad38e-9936-45bf-8e20-b20e5f7958bd', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1009.286951] env[62965]: DEBUG oslo.service.loopingcall [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1009.287186] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1009.287381] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6a9b65d6-bfbd-4bd1-86ff-38bcbde7c032 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.301348] env[62965]: DEBUG nova.compute.manager [None req-7e818b1d-898d-46dc-9cff-8175b4ed1ce6 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1009.301577] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7e818b1d-898d-46dc-9cff-8175b4ed1ce6 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1009.302440] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b6a1a47-670d-4b18-851b-2fb38895f6b5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.309227] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e818b1d-898d-46dc-9cff-8175b4ed1ce6 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1009.309442] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b589af73-8fbc-4729-a334-97e76aa8ced3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.311684] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1009.311684] env[62965]: value = "task-1390255" [ 1009.311684] env[62965]: _type = "Task" [ 1009.311684] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.315728] env[62965]: DEBUG oslo_vmware.api [None req-7e818b1d-898d-46dc-9cff-8175b4ed1ce6 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 1009.315728] env[62965]: value = "task-1390256" [ 1009.315728] env[62965]: _type = "Task" [ 1009.315728] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.321069] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390255, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.325631] env[62965]: DEBUG oslo_vmware.api [None req-7e818b1d-898d-46dc-9cff-8175b4ed1ce6 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390256, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.332160] env[62965]: DEBUG oslo_concurrency.lockutils [req-fbea45f8-c638-4fd9-b825-7f8231818b7d req-cf85fe2d-28f3-4770-a867-f6ba8b70b011 service nova] Releasing lock "refresh_cache-0620c47c-2918-43dc-9e67-f32ca3dd20f1" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.332488] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquired lock "refresh_cache-0620c47c-2918-43dc-9e67-f32ca3dd20f1" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.332638] env[62965]: DEBUG nova.network.neutron [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1009.680619] env[62965]: DEBUG oslo_vmware.api [None req-2b3fc472-7a55-4324-9044-8e770a80c956 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390254, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.726947] env[62965]: DEBUG nova.scheduler.client.report [None req-01458532-419b-4fa3-bf4c-ef681d3f6607 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1009.823100] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390255, 'name': CreateVM_Task, 'duration_secs': 0.350005} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.823693] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1009.824428] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.824601] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.825043] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1009.828285] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-570c8dfa-dc8b-4be8-b068-879f8e685ce3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.830095] env[62965]: DEBUG oslo_vmware.api [None req-7e818b1d-898d-46dc-9cff-8175b4ed1ce6 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390256, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.833633] env[62965]: DEBUG oslo_vmware.api [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1009.833633] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]520a0218-9259-ae7c-595c-5128a7009d5d" [ 1009.833633] env[62965]: _type = "Task" [ 1009.833633] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.847476] env[62965]: DEBUG oslo_vmware.api [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]520a0218-9259-ae7c-595c-5128a7009d5d, 'name': SearchDatastore_Task, 'duration_secs': 0.008862} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.847799] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.848052] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1009.848294] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.848440] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.848643] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1009.848917] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-65be7e6e-7ee1-45a5-9f61-18ff179bb578 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.864100] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1009.864317] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1009.865089] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54fc0526-e948-4706-a845-fd5f87929f2e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.870246] env[62965]: DEBUG oslo_vmware.api [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1009.870246] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52848ba9-5b72-e63b-3b19-98cfd71219ad" [ 1009.870246] env[62965]: _type = "Task" [ 1009.870246] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.878793] env[62965]: DEBUG oslo_vmware.api [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52848ba9-5b72-e63b-3b19-98cfd71219ad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.879552] env[62965]: DEBUG nova.network.neutron [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1010.069383] env[62965]: DEBUG nova.network.neutron [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Updating instance_info_cache with network_info: [{"id": "b178e278-58df-4607-ae3b-1d6c07cd22e7", "address": "fa:16:3e:45:7e:ac", "network": {"id": "a0a5e605-d580-4b7b-b3c0-5c7395bd5a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1679618017-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ed71548c085499981fa1b102d660368", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb178e278-58", "ovs_interfaceid": "b178e278-58df-4607-ae3b-1d6c07cd22e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.181416] env[62965]: DEBUG oslo_vmware.api [None req-2b3fc472-7a55-4324-9044-8e770a80c956 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390254, 'name': ReconfigVM_Task, 'duration_secs': 1.203128} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.182136] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b3fc472-7a55-4324-9044-8e770a80c956 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-295045', 'volume_id': 'bbfeace6-81aa-40a3-8563-95c2c9a1c7ef', 'name': 'volume-bbfeace6-81aa-40a3-8563-95c2c9a1c7ef', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a63b301d-8d52-4fec-951f-a15cda9449ac', 'attached_at': '', 'detached_at': '', 'volume_id': 'bbfeace6-81aa-40a3-8563-95c2c9a1c7ef', 'serial': 'bbfeace6-81aa-40a3-8563-95c2c9a1c7ef'} {{(pid=62965) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1010.326280] env[62965]: DEBUG oslo_vmware.api [None req-7e818b1d-898d-46dc-9cff-8175b4ed1ce6 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390256, 'name': PowerOffVM_Task, 'duration_secs': 0.682451} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.326605] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e818b1d-898d-46dc-9cff-8175b4ed1ce6 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1010.326798] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7e818b1d-898d-46dc-9cff-8175b4ed1ce6 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1010.327085] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-178e4ab6-212e-4a36-94a3-3f55fce4b43e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.380534] env[62965]: DEBUG oslo_vmware.api [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52848ba9-5b72-e63b-3b19-98cfd71219ad, 'name': SearchDatastore_Task, 'duration_secs': 0.007783} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.381330] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2d4aa39-0cf1-4a56-bca2-e6be1be39d38 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.386276] env[62965]: DEBUG oslo_vmware.api [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1010.386276] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52d9f11c-0262-21ae-a9d7-787de65d4839" [ 1010.386276] env[62965]: _type = "Task" [ 1010.386276] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.395010] env[62965]: DEBUG oslo_vmware.api [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52d9f11c-0262-21ae-a9d7-787de65d4839, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.396554] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7e818b1d-898d-46dc-9cff-8175b4ed1ce6 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1010.396777] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7e818b1d-898d-46dc-9cff-8175b4ed1ce6 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Deleting contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1010.396959] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e818b1d-898d-46dc-9cff-8175b4ed1ce6 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Deleting the datastore file [datastore2] a0e9cc1a-bc95-4a44-99c5-aabf85a373f4 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1010.397221] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-082a4214-466b-4178-b6f3-063842de6cb7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.404171] env[62965]: DEBUG oslo_vmware.api [None req-7e818b1d-898d-46dc-9cff-8175b4ed1ce6 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for the task: (returnval){ [ 1010.404171] env[62965]: value = "task-1390258" [ 1010.404171] env[62965]: _type = "Task" [ 1010.404171] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.411831] env[62965]: DEBUG oslo_vmware.api [None req-7e818b1d-898d-46dc-9cff-8175b4ed1ce6 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390258, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.563170] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquiring lock "9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.563405] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.571858] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Releasing lock "refresh_cache-0620c47c-2918-43dc-9e67-f32ca3dd20f1" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.572335] env[62965]: DEBUG nova.compute.manager [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Instance network_info: |[{"id": "b178e278-58df-4607-ae3b-1d6c07cd22e7", "address": "fa:16:3e:45:7e:ac", "network": {"id": "a0a5e605-d580-4b7b-b3c0-5c7395bd5a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1679618017-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ed71548c085499981fa1b102d660368", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb178e278-58", "ovs_interfaceid": "b178e278-58df-4607-ae3b-1d6c07cd22e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1010.573047] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:45:7e:ac', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0cd5d325-3053-407e-a4ee-f627e82a23f9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b178e278-58df-4607-ae3b-1d6c07cd22e7', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1010.580393] env[62965]: DEBUG oslo.service.loopingcall [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1010.581618] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1010.581854] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f186d547-51d6-4797-bacd-33c175efc835 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.600149] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquiring lock "913c0536-5e39-4654-b083-2703a6aed328" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.600278] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "913c0536-5e39-4654-b083-2703a6aed328" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.604768] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1010.604768] env[62965]: value = "task-1390259" [ 1010.604768] env[62965]: _type = "Task" [ 1010.604768] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.613344] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390259, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.725792] env[62965]: DEBUG nova.objects.instance [None req-2b3fc472-7a55-4324-9044-8e770a80c956 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lazy-loading 'flavor' on Instance uuid a63b301d-8d52-4fec-951f-a15cda9449ac {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1010.736306] env[62965]: DEBUG oslo_concurrency.lockutils [None req-01458532-419b-4fa3-bf4c-ef681d3f6607 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.231s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.739670] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.432s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.741084] env[62965]: INFO nova.compute.claims [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1011.595394] env[62965]: DEBUG nova.compute.manager [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1011.597950] env[62965]: DEBUG nova.compute.manager [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1011.606462] env[62965]: DEBUG oslo_vmware.api [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52d9f11c-0262-21ae-a9d7-787de65d4839, 'name': SearchDatastore_Task, 'duration_secs': 0.012491} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.614227] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.614494] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] 8efad776-6cbe-49ab-83dc-a410d8417505/8efad776-6cbe-49ab-83dc-a410d8417505.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1011.615664] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-24d1829a-6fa3-4d50-8ca9-1b52c294592e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.627018] env[62965]: DEBUG oslo_vmware.api [None req-7e818b1d-898d-46dc-9cff-8175b4ed1ce6 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Task: {'id': task-1390258, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.128294} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.627315] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390259, 'name': CreateVM_Task, 'duration_secs': 0.627862} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.628938] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e818b1d-898d-46dc-9cff-8175b4ed1ce6 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1011.629171] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7e818b1d-898d-46dc-9cff-8175b4ed1ce6 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Deleted contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1011.629345] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7e818b1d-898d-46dc-9cff-8175b4ed1ce6 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1011.629515] env[62965]: INFO nova.compute.manager [None req-7e818b1d-898d-46dc-9cff-8175b4ed1ce6 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Took 2.33 seconds to destroy the instance on the hypervisor. [ 1011.629800] env[62965]: DEBUG oslo.service.loopingcall [None req-7e818b1d-898d-46dc-9cff-8175b4ed1ce6 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1011.629884] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1011.630189] env[62965]: DEBUG oslo_vmware.api [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1011.630189] env[62965]: value = "task-1390260" [ 1011.630189] env[62965]: _type = "Task" [ 1011.630189] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.630668] env[62965]: DEBUG nova.compute.manager [-] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1011.630769] env[62965]: DEBUG nova.network.neutron [-] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1011.632721] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.632884] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.633239] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1011.633767] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-33340700-3ad7-44de-bc0a-65f84fae25bb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.648153] env[62965]: DEBUG oslo_vmware.api [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 1011.648153] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5275a44e-4d8a-1c6c-6794-123ca0face11" [ 1011.648153] env[62965]: _type = "Task" [ 1011.648153] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.648921] env[62965]: DEBUG oslo_vmware.api [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390260, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.658172] env[62965]: DEBUG oslo_vmware.api [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5275a44e-4d8a-1c6c-6794-123ca0face11, 'name': SearchDatastore_Task, 'duration_secs': 0.009946} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.658571] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.658892] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1011.659398] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1011.659677] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1011.659905] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1011.660248] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-07b3e9ba-825f-48ff-91b0-93ff962eaa31 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.673407] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1011.673627] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1011.674715] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bfff38b5-85d0-40c0-a23f-0fb19689d667 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.680430] env[62965]: DEBUG oslo_vmware.api [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 1011.680430] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52758055-8b10-bc4e-360b-18a28aeda435" [ 1011.680430] env[62965]: _type = "Task" [ 1011.680430] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.689101] env[62965]: DEBUG oslo_vmware.api [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52758055-8b10-bc4e-360b-18a28aeda435, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.701296] env[62965]: INFO nova.scheduler.client.report [None req-01458532-419b-4fa3-bf4c-ef681d3f6607 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Deleted allocation for migration a802f9c1-2aa9-4613-9e17-47bbb76297fb [ 1012.028552] env[62965]: DEBUG nova.compute.manager [req-21fcf93f-95ca-4062-b1b8-7b233715b907 req-494e101c-b03f-407c-bcca-8afab40d4b3e service nova] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Received event network-vif-deleted-6ffa156f-6a68-424e-b324-660892b9ad22 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1012.028796] env[62965]: INFO nova.compute.manager [req-21fcf93f-95ca-4062-b1b8-7b233715b907 req-494e101c-b03f-407c-bcca-8afab40d4b3e service nova] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Neutron deleted interface 6ffa156f-6a68-424e-b324-660892b9ad22; detaching it from the instance and deleting it from the info cache [ 1012.028935] env[62965]: DEBUG nova.network.neutron [req-21fcf93f-95ca-4062-b1b8-7b233715b907 req-494e101c-b03f-407c-bcca-8afab40d4b3e service nova] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.122291] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2b3fc472-7a55-4324-9044-8e770a80c956 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "a63b301d-8d52-4fec-951f-a15cda9449ac" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.653s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.149469] env[62965]: DEBUG oslo_vmware.api [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390260, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.151070] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.152532] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.191105] env[62965]: DEBUG oslo_vmware.api [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52758055-8b10-bc4e-360b-18a28aeda435, 'name': SearchDatastore_Task, 'duration_secs': 0.009589} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.192046] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b5e3352-be11-4977-b968-d2d21edd12bf {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.197719] env[62965]: DEBUG oslo_vmware.api [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 1012.197719] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]523411c9-612c-45f6-a3c1-c6014e5aebee" [ 1012.197719] env[62965]: _type = "Task" [ 1012.197719] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.207174] env[62965]: DEBUG oslo_concurrency.lockutils [None req-01458532-419b-4fa3-bf4c-ef681d3f6607 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "464a3a31-06bf-4662-a5b1-d185d9a64f0b" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.940s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.208306] env[62965]: DEBUG oslo_vmware.api [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]523411c9-612c-45f6-a3c1-c6014e5aebee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.341455] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9991547d-3478-429e-9398-0d07ca0ddab7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.349245] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6868c657-db90-4ae8-8b92-602a83f4f8a0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.380272] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9b5a0ae-736c-4331-bbb9-569561d17ee1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.388067] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bfff610-8cb9-41e8-bc83-56f52ea60847 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.401590] env[62965]: DEBUG nova.compute.provider_tree [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1012.447670] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7a26068c-4b51-4f97-af71-622aad6bddf0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "464a3a31-06bf-4662-a5b1-d185d9a64f0b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.449332] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7a26068c-4b51-4f97-af71-622aad6bddf0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "464a3a31-06bf-4662-a5b1-d185d9a64f0b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.449332] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7a26068c-4b51-4f97-af71-622aad6bddf0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "464a3a31-06bf-4662-a5b1-d185d9a64f0b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.449332] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7a26068c-4b51-4f97-af71-622aad6bddf0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "464a3a31-06bf-4662-a5b1-d185d9a64f0b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.453223] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7a26068c-4b51-4f97-af71-622aad6bddf0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "464a3a31-06bf-4662-a5b1-d185d9a64f0b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.461026] env[62965]: INFO nova.compute.manager [None req-7a26068c-4b51-4f97-af71-622aad6bddf0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Terminating instance [ 1012.495101] env[62965]: DEBUG nova.network.neutron [-] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.531296] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5420dea3-35d0-47bb-9969-0565d270c2de {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.541924] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9085c19d-c880-4e48-ae65-854f77518b5f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.570784] env[62965]: DEBUG nova.compute.manager [req-21fcf93f-95ca-4062-b1b8-7b233715b907 req-494e101c-b03f-407c-bcca-8afab40d4b3e service nova] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Detach interface failed, port_id=6ffa156f-6a68-424e-b324-660892b9ad22, reason: Instance a0e9cc1a-bc95-4a44-99c5-aabf85a373f4 could not be found. {{(pid=62965) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11294}} [ 1012.644210] env[62965]: DEBUG oslo_vmware.api [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390260, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.587999} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.644526] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] 8efad776-6cbe-49ab-83dc-a410d8417505/8efad776-6cbe-49ab-83dc-a410d8417505.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1012.644706] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1012.644963] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-481689d9-0948-4048-bb13-d3ab6a9b3595 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.652195] env[62965]: DEBUG oslo_vmware.api [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1012.652195] env[62965]: value = "task-1390261" [ 1012.652195] env[62965]: _type = "Task" [ 1012.652195] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.660022] env[62965]: DEBUG oslo_vmware.api [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390261, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.707890] env[62965]: DEBUG oslo_vmware.api [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]523411c9-612c-45f6-a3c1-c6014e5aebee, 'name': SearchDatastore_Task, 'duration_secs': 0.010748} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.708170] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.708432] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] 0620c47c-2918-43dc-9e67-f32ca3dd20f1/0620c47c-2918-43dc-9e67-f32ca3dd20f1.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1012.708686] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8968589b-1dd2-4273-8d6c-9bab2ff38eb0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.715346] env[62965]: DEBUG oslo_vmware.api [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 1012.715346] env[62965]: value = "task-1390262" [ 1012.715346] env[62965]: _type = "Task" [ 1012.715346] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.723136] env[62965]: DEBUG oslo_vmware.api [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390262, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.922880] env[62965]: ERROR nova.scheduler.client.report [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [req-d93852b0-9e42-4387-8750-b212876b4d3c] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-d93852b0-9e42-4387-8750-b212876b4d3c"}]} [ 1012.946316] env[62965]: DEBUG nova.scheduler.client.report [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Refreshing inventories for resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 1012.963606] env[62965]: DEBUG nova.scheduler.client.report [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Updating ProviderTree inventory for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 1012.963844] env[62965]: DEBUG nova.compute.provider_tree [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1012.967473] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3dfc7706-1742-4cc3-8097-297bcd700574 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquiring lock "a63b301d-8d52-4fec-951f-a15cda9449ac" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.967747] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3dfc7706-1742-4cc3-8097-297bcd700574 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "a63b301d-8d52-4fec-951f-a15cda9449ac" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.967966] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3dfc7706-1742-4cc3-8097-297bcd700574 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquiring lock "a63b301d-8d52-4fec-951f-a15cda9449ac-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.968221] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3dfc7706-1742-4cc3-8097-297bcd700574 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "a63b301d-8d52-4fec-951f-a15cda9449ac-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.968415] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3dfc7706-1742-4cc3-8097-297bcd700574 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "a63b301d-8d52-4fec-951f-a15cda9449ac-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.970552] env[62965]: DEBUG nova.compute.manager [None req-7a26068c-4b51-4f97-af71-622aad6bddf0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1012.970809] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7a26068c-4b51-4f97-af71-622aad6bddf0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1012.971409] env[62965]: INFO nova.compute.manager [None req-3dfc7706-1742-4cc3-8097-297bcd700574 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Terminating instance [ 1012.973609] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cf92b47-f90e-4fcb-894f-07a220bb1328 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.978866] env[62965]: DEBUG nova.scheduler.client.report [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Refreshing aggregate associations for resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8, aggregates: None {{(pid=62965) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 1012.986418] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a26068c-4b51-4f97-af71-622aad6bddf0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1012.987335] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4df25f1e-2674-4da2-b1df-e2283d943dad {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.994573] env[62965]: DEBUG oslo_vmware.api [None req-7a26068c-4b51-4f97-af71-622aad6bddf0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 1012.994573] env[62965]: value = "task-1390263" [ 1012.994573] env[62965]: _type = "Task" [ 1012.994573] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.999559] env[62965]: DEBUG nova.scheduler.client.report [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Refreshing trait associations for resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64 {{(pid=62965) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 1013.002849] env[62965]: INFO nova.compute.manager [-] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Took 1.37 seconds to deallocate network for instance. [ 1013.009025] env[62965]: DEBUG oslo_vmware.api [None req-7a26068c-4b51-4f97-af71-622aad6bddf0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390263, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.055192] env[62965]: DEBUG oslo_concurrency.lockutils [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "df03e327-959c-42af-85cd-a7af2ec4d0f7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.055508] env[62965]: DEBUG oslo_concurrency.lockutils [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "df03e327-959c-42af-85cd-a7af2ec4d0f7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.163156] env[62965]: DEBUG oslo_vmware.api [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390261, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.213525} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.166440] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1013.167699] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cfd0152-2184-4a78-a897-c84bf0ab4f34 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.192350] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] 8efad776-6cbe-49ab-83dc-a410d8417505/8efad776-6cbe-49ab-83dc-a410d8417505.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1013.197200] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aca7afb5-164c-427c-a552-eef6d80753c8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.221607] env[62965]: DEBUG oslo_vmware.api [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1013.221607] env[62965]: value = "task-1390264" [ 1013.221607] env[62965]: _type = "Task" [ 1013.221607] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.232451] env[62965]: DEBUG oslo_vmware.api [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390262, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.238841] env[62965]: DEBUG oslo_vmware.api [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390264, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.248316] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f97d07d9-2790-41f3-ad61-8970f48e7075 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.256885] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47fe19de-e693-4428-9c72-c85979addbfe {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.295058] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25f6790b-84b8-45b8-87b7-841ee6f63926 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.301576] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32d0f52f-ef62-4add-9e7e-f96326c74335 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.314953] env[62965]: DEBUG nova.compute.provider_tree [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1013.482856] env[62965]: DEBUG nova.compute.manager [None req-3dfc7706-1742-4cc3-8097-297bcd700574 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1013.483601] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-3dfc7706-1742-4cc3-8097-297bcd700574 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1013.484066] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64e43230-6f50-43dc-9b05-6ecd530fcf9f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.494698] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dfc7706-1742-4cc3-8097-297bcd700574 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1013.494924] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-049cba51-2a8b-4177-b1af-652430db3d6b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.504682] env[62965]: DEBUG oslo_vmware.api [None req-7a26068c-4b51-4f97-af71-622aad6bddf0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390263, 'name': PowerOffVM_Task, 'duration_secs': 0.282294} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.505873] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a26068c-4b51-4f97-af71-622aad6bddf0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1013.506035] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7a26068c-4b51-4f97-af71-622aad6bddf0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1013.506345] env[62965]: DEBUG oslo_vmware.api [None req-3dfc7706-1742-4cc3-8097-297bcd700574 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Waiting for the task: (returnval){ [ 1013.506345] env[62965]: value = "task-1390265" [ 1013.506345] env[62965]: _type = "Task" [ 1013.506345] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.506530] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a63260a9-d6c8-426f-8e1c-37f4ce1d62f4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.515494] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7e818b1d-898d-46dc-9cff-8175b4ed1ce6 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.559068] env[62965]: DEBUG nova.compute.manager [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1013.577228] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7a26068c-4b51-4f97-af71-622aad6bddf0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1013.577648] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7a26068c-4b51-4f97-af71-622aad6bddf0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Deleting contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1013.577988] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a26068c-4b51-4f97-af71-622aad6bddf0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Deleting the datastore file [datastore1] 464a3a31-06bf-4662-a5b1-d185d9a64f0b {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1013.578435] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cc247dd8-318c-4c26-91d1-df89e47fae02 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.589019] env[62965]: DEBUG oslo_vmware.api [None req-7a26068c-4b51-4f97-af71-622aad6bddf0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 1013.589019] env[62965]: value = "task-1390267" [ 1013.589019] env[62965]: _type = "Task" [ 1013.589019] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.594806] env[62965]: DEBUG oslo_vmware.api [None req-7a26068c-4b51-4f97-af71-622aad6bddf0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390267, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.725791] env[62965]: DEBUG oslo_vmware.api [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390262, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.578503} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.728901] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] 0620c47c-2918-43dc-9e67-f32ca3dd20f1/0620c47c-2918-43dc-9e67-f32ca3dd20f1.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1013.729139] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1013.729892] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-97a92562-8568-456a-8fc8-d47c94d540a1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.736455] env[62965]: DEBUG oslo_vmware.api [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390264, 'name': ReconfigVM_Task, 'duration_secs': 0.317332} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.737582] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Reconfigured VM instance instance-0000005f to attach disk [datastore1] 8efad776-6cbe-49ab-83dc-a410d8417505/8efad776-6cbe-49ab-83dc-a410d8417505.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1013.738279] env[62965]: DEBUG oslo_vmware.api [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 1013.738279] env[62965]: value = "task-1390268" [ 1013.738279] env[62965]: _type = "Task" [ 1013.738279] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.738474] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1878f0ae-daaa-4d8e-b247-b2ea44973a19 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.747615] env[62965]: DEBUG oslo_vmware.api [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390268, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.748814] env[62965]: DEBUG oslo_vmware.api [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1013.748814] env[62965]: value = "task-1390269" [ 1013.748814] env[62965]: _type = "Task" [ 1013.748814] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.757021] env[62965]: DEBUG oslo_vmware.api [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390269, 'name': Rename_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.848549] env[62965]: DEBUG nova.scheduler.client.report [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Updated inventory for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with generation 123 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 1013.848826] env[62965]: DEBUG nova.compute.provider_tree [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Updating resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 generation from 123 to 124 during operation: update_inventory {{(pid=62965) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1013.849011] env[62965]: DEBUG nova.compute.provider_tree [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1014.017723] env[62965]: DEBUG oslo_vmware.api [None req-3dfc7706-1742-4cc3-8097-297bcd700574 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390265, 'name': PowerOffVM_Task, 'duration_secs': 0.204335} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.017999] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dfc7706-1742-4cc3-8097-297bcd700574 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1014.018188] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-3dfc7706-1742-4cc3-8097-297bcd700574 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1014.018445] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f815580d-06e8-4d5b-8fd1-1bbd87ddfdd5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.028337] env[62965]: DEBUG oslo_concurrency.lockutils [None req-6a2cec05-0481-43a8-82f9-6d82770f9c14 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquiring lock "b59d6023-a047-4b32-bd18-1c7cc05f67d7" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.028577] env[62965]: DEBUG oslo_concurrency.lockutils [None req-6a2cec05-0481-43a8-82f9-6d82770f9c14 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "b59d6023-a047-4b32-bd18-1c7cc05f67d7" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.080267] env[62965]: DEBUG oslo_concurrency.lockutils [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.096293] env[62965]: DEBUG oslo_vmware.api [None req-7a26068c-4b51-4f97-af71-622aad6bddf0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390267, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145061} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.096293] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a26068c-4b51-4f97-af71-622aad6bddf0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1014.096293] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7a26068c-4b51-4f97-af71-622aad6bddf0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Deleted contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1014.096293] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-7a26068c-4b51-4f97-af71-622aad6bddf0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1014.096494] env[62965]: INFO nova.compute.manager [None req-7a26068c-4b51-4f97-af71-622aad6bddf0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1014.096766] env[62965]: DEBUG oslo.service.loopingcall [None req-7a26068c-4b51-4f97-af71-622aad6bddf0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1014.096972] env[62965]: DEBUG nova.compute.manager [-] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1014.097113] env[62965]: DEBUG nova.network.neutron [-] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1014.099237] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-3dfc7706-1742-4cc3-8097-297bcd700574 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1014.099433] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-3dfc7706-1742-4cc3-8097-297bcd700574 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Deleting contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1014.099611] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-3dfc7706-1742-4cc3-8097-297bcd700574 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Deleting the datastore file [datastore2] a63b301d-8d52-4fec-951f-a15cda9449ac {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1014.100837] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5d27b615-6b3c-43e1-850c-fb48fdc4f268 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.105996] env[62965]: DEBUG oslo_vmware.api [None req-3dfc7706-1742-4cc3-8097-297bcd700574 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Waiting for the task: (returnval){ [ 1014.105996] env[62965]: value = "task-1390271" [ 1014.105996] env[62965]: _type = "Task" [ 1014.105996] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.113238] env[62965]: DEBUG oslo_vmware.api [None req-3dfc7706-1742-4cc3-8097-297bcd700574 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390271, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.256318] env[62965]: DEBUG oslo_vmware.api [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390268, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063944} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.257066] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1014.257890] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c826dc32-02ae-49cb-883f-1cf5bf76e451 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.263472] env[62965]: DEBUG oslo_vmware.api [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390269, 'name': Rename_Task, 'duration_secs': 0.129756} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.264043] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1014.264295] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fe891252-ba10-4444-a283-9386be8dd0a4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.283359] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Reconfiguring VM instance instance-00000060 to attach disk [datastore1] 0620c47c-2918-43dc-9e67-f32ca3dd20f1/0620c47c-2918-43dc-9e67-f32ca3dd20f1.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1014.284126] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6ee60bcc-32e5-456f-b44a-728d52a16ce7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.300792] env[62965]: DEBUG oslo_vmware.api [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1014.300792] env[62965]: value = "task-1390272" [ 1014.300792] env[62965]: _type = "Task" [ 1014.300792] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.306171] env[62965]: DEBUG oslo_vmware.api [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 1014.306171] env[62965]: value = "task-1390273" [ 1014.306171] env[62965]: _type = "Task" [ 1014.306171] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.313958] env[62965]: DEBUG oslo_vmware.api [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390272, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.317426] env[62965]: DEBUG oslo_vmware.api [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390273, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.351038] env[62965]: DEBUG nova.compute.manager [req-38b5dbb2-7f58-488c-ac89-dbda6cbd487c req-ee616c8b-2e57-47f7-991e-c3a9fe48f868 service nova] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Received event network-vif-deleted-170ea80e-f3c0-4647-af9d-34ffed538c33 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1014.351198] env[62965]: INFO nova.compute.manager [req-38b5dbb2-7f58-488c-ac89-dbda6cbd487c req-ee616c8b-2e57-47f7-991e-c3a9fe48f868 service nova] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Neutron deleted interface 170ea80e-f3c0-4647-af9d-34ffed538c33; detaching it from the instance and deleting it from the info cache [ 1014.351360] env[62965]: DEBUG nova.network.neutron [req-38b5dbb2-7f58-488c-ac89-dbda6cbd487c req-ee616c8b-2e57-47f7-991e-c3a9fe48f868 service nova] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.354011] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.614s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.354481] env[62965]: DEBUG nova.compute.manager [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1014.357576] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.207s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.359105] env[62965]: INFO nova.compute.claims [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1014.531727] env[62965]: INFO nova.compute.manager [None req-6a2cec05-0481-43a8-82f9-6d82770f9c14 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Detaching volume 437ab7f8-7ad9-4d87-8f07-900702627943 [ 1014.568709] env[62965]: INFO nova.virt.block_device [None req-6a2cec05-0481-43a8-82f9-6d82770f9c14 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Attempting to driver detach volume 437ab7f8-7ad9-4d87-8f07-900702627943 from mountpoint /dev/sdb [ 1014.568987] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a2cec05-0481-43a8-82f9-6d82770f9c14 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Volume detach. Driver type: vmdk {{(pid=62965) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1014.569214] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a2cec05-0481-43a8-82f9-6d82770f9c14 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-295026', 'volume_id': '437ab7f8-7ad9-4d87-8f07-900702627943', 'name': 'volume-437ab7f8-7ad9-4d87-8f07-900702627943', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'b59d6023-a047-4b32-bd18-1c7cc05f67d7', 'attached_at': '', 'detached_at': '', 'volume_id': '437ab7f8-7ad9-4d87-8f07-900702627943', 'serial': '437ab7f8-7ad9-4d87-8f07-900702627943'} {{(pid=62965) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1014.570149] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-782b8705-4317-42a1-850b-bc113a93ae08 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.592044] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf91d47e-50f4-46c0-8e31-f324f80e9d87 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.599110] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7c64e09-e06d-489b-8bf2-75a5b2215f5c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.624899] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2577e16f-fa3d-4315-80bd-77ca53801d29 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.632337] env[62965]: DEBUG oslo_vmware.api [None req-3dfc7706-1742-4cc3-8097-297bcd700574 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Task: {'id': task-1390271, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.142493} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.643701] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-3dfc7706-1742-4cc3-8097-297bcd700574 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1014.643701] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-3dfc7706-1742-4cc3-8097-297bcd700574 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Deleted contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1014.643701] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-3dfc7706-1742-4cc3-8097-297bcd700574 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1014.643701] env[62965]: INFO nova.compute.manager [None req-3dfc7706-1742-4cc3-8097-297bcd700574 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1014.643701] env[62965]: DEBUG oslo.service.loopingcall [None req-3dfc7706-1742-4cc3-8097-297bcd700574 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1014.643943] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a2cec05-0481-43a8-82f9-6d82770f9c14 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] The volume has not been displaced from its original location: [datastore1] volume-437ab7f8-7ad9-4d87-8f07-900702627943/volume-437ab7f8-7ad9-4d87-8f07-900702627943.vmdk. No consolidation needed. {{(pid=62965) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1014.649819] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a2cec05-0481-43a8-82f9-6d82770f9c14 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Reconfiguring VM instance instance-00000048 to detach disk 2001 {{(pid=62965) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1014.649819] env[62965]: DEBUG nova.compute.manager [-] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1014.649819] env[62965]: DEBUG nova.network.neutron [-] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1014.651556] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b32af265-a63d-45d5-a347-ea582512ca69 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.670237] env[62965]: DEBUG oslo_vmware.api [None req-6a2cec05-0481-43a8-82f9-6d82770f9c14 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 1014.670237] env[62965]: value = "task-1390274" [ 1014.670237] env[62965]: _type = "Task" [ 1014.670237] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.677858] env[62965]: DEBUG oslo_vmware.api [None req-6a2cec05-0481-43a8-82f9-6d82770f9c14 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390274, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.813580] env[62965]: DEBUG oslo_vmware.api [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390272, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.819194] env[62965]: DEBUG oslo_vmware.api [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390273, 'name': ReconfigVM_Task, 'duration_secs': 0.451872} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.822299] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Reconfigured VM instance instance-00000060 to attach disk [datastore1] 0620c47c-2918-43dc-9e67-f32ca3dd20f1/0620c47c-2918-43dc-9e67-f32ca3dd20f1.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1014.822299] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-233eae31-f0df-4cf9-a516-01654a6dcef2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.829135] env[62965]: DEBUG oslo_vmware.api [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 1014.829135] env[62965]: value = "task-1390275" [ 1014.829135] env[62965]: _type = "Task" [ 1014.829135] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.830735] env[62965]: DEBUG nova.network.neutron [-] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.838347] env[62965]: DEBUG oslo_vmware.api [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390275, 'name': Rename_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.853955] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-488ba968-4ff5-47ca-8686-3e9c5fce7bdc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.859794] env[62965]: DEBUG nova.compute.utils [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1014.862385] env[62965]: DEBUG nova.compute.manager [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1014.862657] env[62965]: DEBUG nova.network.neutron [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1014.869646] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4515f6f1-f064-436e-a2ee-ee9d2841da34 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.901507] env[62965]: DEBUG nova.compute.manager [req-38b5dbb2-7f58-488c-ac89-dbda6cbd487c req-ee616c8b-2e57-47f7-991e-c3a9fe48f868 service nova] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Detach interface failed, port_id=170ea80e-f3c0-4647-af9d-34ffed538c33, reason: Instance 464a3a31-06bf-4662-a5b1-d185d9a64f0b could not be found. {{(pid=62965) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11294}} [ 1014.948044] env[62965]: DEBUG nova.policy [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03c0a9a0ec3c46bf883c68d86cb816d6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '978666785b0743bb843de96585f96fb4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 1015.070194] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-051764b8-dc8a-41e8-a19d-da9e61345f43 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.078483] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8309bff-b69f-44fd-8692-b42441f71c37 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.109389] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-853d5160-b00e-4bc5-869b-1c44a281ad52 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.117056] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25dc9db7-c3b3-4197-9488-2c6e756ba11c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.130920] env[62965]: DEBUG nova.compute.provider_tree [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1015.179941] env[62965]: DEBUG oslo_vmware.api [None req-6a2cec05-0481-43a8-82f9-6d82770f9c14 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390274, 'name': ReconfigVM_Task, 'duration_secs': 0.215656} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.180264] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a2cec05-0481-43a8-82f9-6d82770f9c14 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Reconfigured VM instance instance-00000048 to detach disk 2001 {{(pid=62965) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1015.184958] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-148da859-29e0-4345-9482-75f18a18b6ad {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.202576] env[62965]: DEBUG oslo_vmware.api [None req-6a2cec05-0481-43a8-82f9-6d82770f9c14 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 1015.202576] env[62965]: value = "task-1390276" [ 1015.202576] env[62965]: _type = "Task" [ 1015.202576] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.210783] env[62965]: DEBUG oslo_vmware.api [None req-6a2cec05-0481-43a8-82f9-6d82770f9c14 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390276, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.218758] env[62965]: DEBUG nova.network.neutron [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Successfully created port: 0870a7fe-8ea7-42dd-bbe4-558f7f5de034 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1015.313451] env[62965]: DEBUG oslo_vmware.api [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390272, 'name': PowerOnVM_Task, 'duration_secs': 0.526248} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.313753] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1015.313974] env[62965]: DEBUG nova.compute.manager [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1015.314770] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-149611b1-f1a9-4c1c-88ae-d4c7dc2aa91d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.336165] env[62965]: INFO nova.compute.manager [-] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Took 1.24 seconds to deallocate network for instance. [ 1015.336531] env[62965]: DEBUG oslo_vmware.api [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390275, 'name': Rename_Task, 'duration_secs': 0.137757} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.339231] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1015.339231] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c3448cd6-1747-4ff1-a54d-bf374f73d646 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.347132] env[62965]: DEBUG oslo_vmware.api [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 1015.347132] env[62965]: value = "task-1390277" [ 1015.347132] env[62965]: _type = "Task" [ 1015.347132] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.356033] env[62965]: DEBUG oslo_vmware.api [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390277, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.368047] env[62965]: DEBUG nova.compute.manager [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1015.634335] env[62965]: DEBUG nova.scheduler.client.report [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1015.710185] env[62965]: DEBUG oslo_vmware.api [None req-6a2cec05-0481-43a8-82f9-6d82770f9c14 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390276, 'name': ReconfigVM_Task, 'duration_secs': 0.140407} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.710517] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-6a2cec05-0481-43a8-82f9-6d82770f9c14 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-295026', 'volume_id': '437ab7f8-7ad9-4d87-8f07-900702627943', 'name': 'volume-437ab7f8-7ad9-4d87-8f07-900702627943', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': 'b59d6023-a047-4b32-bd18-1c7cc05f67d7', 'attached_at': '', 'detached_at': '', 'volume_id': '437ab7f8-7ad9-4d87-8f07-900702627943', 'serial': '437ab7f8-7ad9-4d87-8f07-900702627943'} {{(pid=62965) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1015.738528] env[62965]: DEBUG nova.network.neutron [-] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.832103] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.845321] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7a26068c-4b51-4f97-af71-622aad6bddf0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.857301] env[62965]: DEBUG oslo_vmware.api [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390277, 'name': PowerOnVM_Task, 'duration_secs': 0.430628} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.857575] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1015.857813] env[62965]: INFO nova.compute.manager [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Took 8.85 seconds to spawn the instance on the hypervisor. [ 1015.858030] env[62965]: DEBUG nova.compute.manager [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1015.858798] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7402e932-13fa-47be-a412-b1a30a45c402 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.139382] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.781s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.139729] env[62965]: DEBUG nova.compute.manager [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1016.143046] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.991s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.144447] env[62965]: INFO nova.compute.claims [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1016.241334] env[62965]: INFO nova.compute.manager [-] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Took 1.59 seconds to deallocate network for instance. [ 1016.254618] env[62965]: DEBUG nova.objects.instance [None req-6a2cec05-0481-43a8-82f9-6d82770f9c14 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lazy-loading 'flavor' on Instance uuid b59d6023-a047-4b32-bd18-1c7cc05f67d7 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1016.374903] env[62965]: INFO nova.compute.manager [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Took 13.72 seconds to build instance. [ 1016.377934] env[62965]: DEBUG nova.compute.manager [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1016.386838] env[62965]: DEBUG nova.compute.manager [req-ec860d03-8a54-47da-993a-fa0d6d0b6b2e req-14b8f7bb-a871-41e9-82dd-81e10bfcf48d service nova] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Received event network-vif-deleted-219e79ce-5ab0-49eb-a38f-6b1ba814cdf3 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1016.402429] env[62965]: DEBUG nova.virt.hardware [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1016.402729] env[62965]: DEBUG nova.virt.hardware [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1016.402899] env[62965]: DEBUG nova.virt.hardware [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1016.403115] env[62965]: DEBUG nova.virt.hardware [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1016.403250] env[62965]: DEBUG nova.virt.hardware [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1016.403395] env[62965]: DEBUG nova.virt.hardware [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1016.403602] env[62965]: DEBUG nova.virt.hardware [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1016.403764] env[62965]: DEBUG nova.virt.hardware [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1016.403933] env[62965]: DEBUG nova.virt.hardware [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1016.404111] env[62965]: DEBUG nova.virt.hardware [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1016.404292] env[62965]: DEBUG nova.virt.hardware [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1016.405543] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e731fee6-7bf5-4532-a8ef-12c95ccf4b63 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.413949] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e7e3746-e695-4ee0-8463-1261debdabf3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.633776] env[62965]: DEBUG nova.compute.manager [req-b5544872-dc52-4d07-99ec-0b8f75593afd req-f36c42c8-cef6-4c3a-b48c-3b8f2f50a1cd service nova] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Received event network-vif-plugged-0870a7fe-8ea7-42dd-bbe4-558f7f5de034 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1016.633997] env[62965]: DEBUG oslo_concurrency.lockutils [req-b5544872-dc52-4d07-99ec-0b8f75593afd req-f36c42c8-cef6-4c3a-b48c-3b8f2f50a1cd service nova] Acquiring lock "8ae1fde5-e94e-43ed-816a-9e22633dfdd1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.634275] env[62965]: DEBUG oslo_concurrency.lockutils [req-b5544872-dc52-4d07-99ec-0b8f75593afd req-f36c42c8-cef6-4c3a-b48c-3b8f2f50a1cd service nova] Lock "8ae1fde5-e94e-43ed-816a-9e22633dfdd1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.634454] env[62965]: DEBUG oslo_concurrency.lockutils [req-b5544872-dc52-4d07-99ec-0b8f75593afd req-f36c42c8-cef6-4c3a-b48c-3b8f2f50a1cd service nova] Lock "8ae1fde5-e94e-43ed-816a-9e22633dfdd1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.634630] env[62965]: DEBUG nova.compute.manager [req-b5544872-dc52-4d07-99ec-0b8f75593afd req-f36c42c8-cef6-4c3a-b48c-3b8f2f50a1cd service nova] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] No waiting events found dispatching network-vif-plugged-0870a7fe-8ea7-42dd-bbe4-558f7f5de034 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1016.634799] env[62965]: WARNING nova.compute.manager [req-b5544872-dc52-4d07-99ec-0b8f75593afd req-f36c42c8-cef6-4c3a-b48c-3b8f2f50a1cd service nova] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Received unexpected event network-vif-plugged-0870a7fe-8ea7-42dd-bbe4-558f7f5de034 for instance with vm_state building and task_state spawning. [ 1016.644071] env[62965]: DEBUG nova.compute.utils [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1016.649207] env[62965]: DEBUG nova.compute.manager [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1016.649412] env[62965]: DEBUG nova.network.neutron [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1016.689431] env[62965]: DEBUG nova.policy [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e7510e1ec5ee4253b67be988b6b5ab51', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5d2d71b59ede430f846520bf49249e4a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 1016.725566] env[62965]: DEBUG nova.network.neutron [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Successfully updated port: 0870a7fe-8ea7-42dd-bbe4-558f7f5de034 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1016.748460] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3dfc7706-1742-4cc3-8097-297bcd700574 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.866592] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e6919be8-1742-47e5-8e55-6cb95da9c828 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "8efad776-6cbe-49ab-83dc-a410d8417505" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.866881] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e6919be8-1742-47e5-8e55-6cb95da9c828 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "8efad776-6cbe-49ab-83dc-a410d8417505" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.867122] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e6919be8-1742-47e5-8e55-6cb95da9c828 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "8efad776-6cbe-49ab-83dc-a410d8417505-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.867319] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e6919be8-1742-47e5-8e55-6cb95da9c828 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "8efad776-6cbe-49ab-83dc-a410d8417505-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.867495] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e6919be8-1742-47e5-8e55-6cb95da9c828 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "8efad776-6cbe-49ab-83dc-a410d8417505-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.870937] env[62965]: INFO nova.compute.manager [None req-e6919be8-1742-47e5-8e55-6cb95da9c828 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Terminating instance [ 1016.877709] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2483cf57-aaaa-404d-b563-9514ccf770ac tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "0620c47c-2918-43dc-9e67-f32ca3dd20f1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.236s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.926121] env[62965]: DEBUG nova.network.neutron [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Successfully created port: 9f81f2ea-9520-4f9c-a2f4-e6c3b1ab1216 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1017.153924] env[62965]: DEBUG nova.compute.manager [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1017.228210] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "refresh_cache-8ae1fde5-e94e-43ed-816a-9e22633dfdd1" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1017.228359] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquired lock "refresh_cache-8ae1fde5-e94e-43ed-816a-9e22633dfdd1" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1017.228508] env[62965]: DEBUG nova.network.neutron [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1017.261115] env[62965]: DEBUG oslo_concurrency.lockutils [None req-6a2cec05-0481-43a8-82f9-6d82770f9c14 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "b59d6023-a047-4b32-bd18-1c7cc05f67d7" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.232s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.335336] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de6cfd96-051b-4647-af8b-9e59d96a78a0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.343655] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff41c290-fc32-427d-9875-de82bcf839c2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.373734] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-736ac7e9-b4a2-44f3-8d1c-3db62549fc16 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.376725] env[62965]: DEBUG nova.compute.manager [None req-e6919be8-1742-47e5-8e55-6cb95da9c828 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1017.377124] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e6919be8-1742-47e5-8e55-6cb95da9c828 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1017.377672] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a855b09-7875-4761-9479-535882ce832a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.388319] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31322573-cd6b-4dce-9768-9ce4bee2daa3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.393072] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6919be8-1742-47e5-8e55-6cb95da9c828 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1017.393072] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-15f7ecec-08aa-482e-a76e-87679cbcf1c2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.397768] env[62965]: DEBUG oslo_vmware.api [None req-e6919be8-1742-47e5-8e55-6cb95da9c828 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1017.397768] env[62965]: value = "task-1390278" [ 1017.397768] env[62965]: _type = "Task" [ 1017.397768] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.405514] env[62965]: DEBUG nova.compute.provider_tree [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1017.415430] env[62965]: DEBUG oslo_vmware.api [None req-e6919be8-1742-47e5-8e55-6cb95da9c828 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390278, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.759126] env[62965]: DEBUG nova.network.neutron [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1017.889530] env[62965]: DEBUG nova.network.neutron [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Updating instance_info_cache with network_info: [{"id": "0870a7fe-8ea7-42dd-bbe4-558f7f5de034", "address": "fa:16:3e:51:a1:8f", "network": {"id": "62052700-1781-4933-93c8-fa1aafdb0fd2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1931798228-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "978666785b0743bb843de96585f96fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0870a7fe-8e", "ovs_interfaceid": "0870a7fe-8ea7-42dd-bbe4-558f7f5de034", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.909310] env[62965]: DEBUG nova.scheduler.client.report [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1017.917835] env[62965]: DEBUG oslo_vmware.api [None req-e6919be8-1742-47e5-8e55-6cb95da9c828 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390278, 'name': PowerOffVM_Task, 'duration_secs': 0.1707} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.918115] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6919be8-1742-47e5-8e55-6cb95da9c828 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1017.918298] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e6919be8-1742-47e5-8e55-6cb95da9c828 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1017.918544] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9f3e5391-e363-4903-9d98-8dce5b339509 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.921718] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f0be609a-e4c5-4ee3-9430-9d7a07622409 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "0620c47c-2918-43dc-9e67-f32ca3dd20f1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.921961] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f0be609a-e4c5-4ee3-9430-9d7a07622409 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "0620c47c-2918-43dc-9e67-f32ca3dd20f1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.922192] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f0be609a-e4c5-4ee3-9430-9d7a07622409 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "0620c47c-2918-43dc-9e67-f32ca3dd20f1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.922387] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f0be609a-e4c5-4ee3-9430-9d7a07622409 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "0620c47c-2918-43dc-9e67-f32ca3dd20f1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.922554] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f0be609a-e4c5-4ee3-9430-9d7a07622409 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "0620c47c-2918-43dc-9e67-f32ca3dd20f1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.924775] env[62965]: INFO nova.compute.manager [None req-f0be609a-e4c5-4ee3-9430-9d7a07622409 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Terminating instance [ 1018.050354] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e6919be8-1742-47e5-8e55-6cb95da9c828 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1018.050594] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e6919be8-1742-47e5-8e55-6cb95da9c828 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Deleting contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1018.050855] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6919be8-1742-47e5-8e55-6cb95da9c828 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Deleting the datastore file [datastore1] 8efad776-6cbe-49ab-83dc-a410d8417505 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1018.051150] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-815052cd-e9a2-4650-91fa-b54e04d2564c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.057654] env[62965]: DEBUG oslo_vmware.api [None req-e6919be8-1742-47e5-8e55-6cb95da9c828 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1018.057654] env[62965]: value = "task-1390280" [ 1018.057654] env[62965]: _type = "Task" [ 1018.057654] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.065375] env[62965]: DEBUG oslo_vmware.api [None req-e6919be8-1742-47e5-8e55-6cb95da9c828 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390280, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.164854] env[62965]: DEBUG nova.compute.manager [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1018.190240] env[62965]: DEBUG nova.virt.hardware [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1018.190529] env[62965]: DEBUG nova.virt.hardware [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1018.190656] env[62965]: DEBUG nova.virt.hardware [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1018.190914] env[62965]: DEBUG nova.virt.hardware [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1018.191040] env[62965]: DEBUG nova.virt.hardware [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1018.191202] env[62965]: DEBUG nova.virt.hardware [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1018.191413] env[62965]: DEBUG nova.virt.hardware [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1018.191577] env[62965]: DEBUG nova.virt.hardware [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1018.191750] env[62965]: DEBUG nova.virt.hardware [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1018.191917] env[62965]: DEBUG nova.virt.hardware [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1018.192108] env[62965]: DEBUG nova.virt.hardware [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1018.193032] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a23ec59d-ed60-44b2-af6a-dcc95dfe7371 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.201539] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ce9a5e8-bcf0-4e23-89d7-a6c470a3b8ca {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.290032] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquiring lock "b59d6023-a047-4b32-bd18-1c7cc05f67d7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.290398] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "b59d6023-a047-4b32-bd18-1c7cc05f67d7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.290696] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquiring lock "b59d6023-a047-4b32-bd18-1c7cc05f67d7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.290926] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "b59d6023-a047-4b32-bd18-1c7cc05f67d7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.291152] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "b59d6023-a047-4b32-bd18-1c7cc05f67d7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.293769] env[62965]: INFO nova.compute.manager [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Terminating instance [ 1018.392036] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Releasing lock "refresh_cache-8ae1fde5-e94e-43ed-816a-9e22633dfdd1" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1018.392401] env[62965]: DEBUG nova.compute.manager [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Instance network_info: |[{"id": "0870a7fe-8ea7-42dd-bbe4-558f7f5de034", "address": "fa:16:3e:51:a1:8f", "network": {"id": "62052700-1781-4933-93c8-fa1aafdb0fd2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1931798228-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "978666785b0743bb843de96585f96fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0870a7fe-8e", "ovs_interfaceid": "0870a7fe-8ea7-42dd-bbe4-558f7f5de034", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1018.392853] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:51:a1:8f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a10c88d7-d13f-44fd-acee-7a734eb5f56a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0870a7fe-8ea7-42dd-bbe4-558f7f5de034', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1018.400840] env[62965]: DEBUG oslo.service.loopingcall [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1018.401111] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1018.401354] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-37ef090d-ab10-4d37-bb41-bcd3d3db6c21 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.417131] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.274s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.417798] env[62965]: DEBUG nova.compute.manager [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1018.420719] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7e818b1d-898d-46dc-9cff-8175b4ed1ce6 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.905s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.421260] env[62965]: DEBUG nova.objects.instance [None req-7e818b1d-898d-46dc-9cff-8175b4ed1ce6 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lazy-loading 'resources' on Instance uuid a0e9cc1a-bc95-4a44-99c5-aabf85a373f4 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1018.429829] env[62965]: DEBUG nova.compute.manager [None req-f0be609a-e4c5-4ee3-9430-9d7a07622409 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1018.430084] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-f0be609a-e4c5-4ee3-9430-9d7a07622409 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1018.430264] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1018.430264] env[62965]: value = "task-1390281" [ 1018.430264] env[62965]: _type = "Task" [ 1018.430264] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.431718] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c9008db-c2c8-4c88-b071-b0af43e15c9e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.446145] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0be609a-e4c5-4ee3-9430-9d7a07622409 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1018.446145] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390281, 'name': CreateVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.447195] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1d23b8f7-d6cd-4cd6-86f6-10c63fd32202 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.449674] env[62965]: DEBUG nova.compute.manager [req-acec4fcd-d143-44a7-8e6f-25e4265a2a3b req-62f66aad-ce52-4f3e-8082-a3808b3f55a1 service nova] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Received event network-vif-plugged-9f81f2ea-9520-4f9c-a2f4-e6c3b1ab1216 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1018.449895] env[62965]: DEBUG oslo_concurrency.lockutils [req-acec4fcd-d143-44a7-8e6f-25e4265a2a3b req-62f66aad-ce52-4f3e-8082-a3808b3f55a1 service nova] Acquiring lock "9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.450104] env[62965]: DEBUG oslo_concurrency.lockutils [req-acec4fcd-d143-44a7-8e6f-25e4265a2a3b req-62f66aad-ce52-4f3e-8082-a3808b3f55a1 service nova] Lock "9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.450280] env[62965]: DEBUG oslo_concurrency.lockutils [req-acec4fcd-d143-44a7-8e6f-25e4265a2a3b req-62f66aad-ce52-4f3e-8082-a3808b3f55a1 service nova] Lock "9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.450457] env[62965]: DEBUG nova.compute.manager [req-acec4fcd-d143-44a7-8e6f-25e4265a2a3b req-62f66aad-ce52-4f3e-8082-a3808b3f55a1 service nova] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] No waiting events found dispatching network-vif-plugged-9f81f2ea-9520-4f9c-a2f4-e6c3b1ab1216 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1018.450825] env[62965]: WARNING nova.compute.manager [req-acec4fcd-d143-44a7-8e6f-25e4265a2a3b req-62f66aad-ce52-4f3e-8082-a3808b3f55a1 service nova] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Received unexpected event network-vif-plugged-9f81f2ea-9520-4f9c-a2f4-e6c3b1ab1216 for instance with vm_state building and task_state spawning. [ 1018.455971] env[62965]: DEBUG oslo_vmware.api [None req-f0be609a-e4c5-4ee3-9430-9d7a07622409 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 1018.455971] env[62965]: value = "task-1390282" [ 1018.455971] env[62965]: _type = "Task" [ 1018.455971] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.467011] env[62965]: DEBUG oslo_vmware.api [None req-f0be609a-e4c5-4ee3-9430-9d7a07622409 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390282, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.547381] env[62965]: DEBUG nova.network.neutron [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Successfully updated port: 9f81f2ea-9520-4f9c-a2f4-e6c3b1ab1216 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1018.568605] env[62965]: DEBUG oslo_vmware.api [None req-e6919be8-1742-47e5-8e55-6cb95da9c828 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390280, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147702} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.568894] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6919be8-1742-47e5-8e55-6cb95da9c828 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1018.569103] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e6919be8-1742-47e5-8e55-6cb95da9c828 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Deleted contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1018.569286] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e6919be8-1742-47e5-8e55-6cb95da9c828 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1018.569477] env[62965]: INFO nova.compute.manager [None req-e6919be8-1742-47e5-8e55-6cb95da9c828 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Took 1.19 seconds to destroy the instance on the hypervisor. [ 1018.569698] env[62965]: DEBUG oslo.service.loopingcall [None req-e6919be8-1742-47e5-8e55-6cb95da9c828 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1018.569901] env[62965]: DEBUG nova.compute.manager [-] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1018.570065] env[62965]: DEBUG nova.network.neutron [-] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1018.799232] env[62965]: DEBUG nova.compute.manager [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1018.799232] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1018.800444] env[62965]: DEBUG nova.compute.manager [req-2f8ad426-89ab-4ddc-a0af-814bac57d4ea req-8a7cbccb-671c-49c9-a147-53a02b2a6037 service nova] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Received event network-changed-0870a7fe-8ea7-42dd-bbe4-558f7f5de034 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1018.800621] env[62965]: DEBUG nova.compute.manager [req-2f8ad426-89ab-4ddc-a0af-814bac57d4ea req-8a7cbccb-671c-49c9-a147-53a02b2a6037 service nova] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Refreshing instance network info cache due to event network-changed-0870a7fe-8ea7-42dd-bbe4-558f7f5de034. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 1018.801013] env[62965]: DEBUG oslo_concurrency.lockutils [req-2f8ad426-89ab-4ddc-a0af-814bac57d4ea req-8a7cbccb-671c-49c9-a147-53a02b2a6037 service nova] Acquiring lock "refresh_cache-8ae1fde5-e94e-43ed-816a-9e22633dfdd1" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.801210] env[62965]: DEBUG oslo_concurrency.lockutils [req-2f8ad426-89ab-4ddc-a0af-814bac57d4ea req-8a7cbccb-671c-49c9-a147-53a02b2a6037 service nova] Acquired lock "refresh_cache-8ae1fde5-e94e-43ed-816a-9e22633dfdd1" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.801389] env[62965]: DEBUG nova.network.neutron [req-2f8ad426-89ab-4ddc-a0af-814bac57d4ea req-8a7cbccb-671c-49c9-a147-53a02b2a6037 service nova] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Refreshing network info cache for port 0870a7fe-8ea7-42dd-bbe4-558f7f5de034 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1018.803495] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92055d9f-5bbe-4e12-9d49-bb6cb06a79ec {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.811793] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1018.812590] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d94ec827-95b1-4282-9637-5f3aefa158ac {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.819420] env[62965]: DEBUG oslo_vmware.api [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 1018.819420] env[62965]: value = "task-1390283" [ 1018.819420] env[62965]: _type = "Task" [ 1018.819420] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.828218] env[62965]: DEBUG oslo_vmware.api [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390283, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.924281] env[62965]: DEBUG nova.compute.utils [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1018.925704] env[62965]: DEBUG nova.compute.manager [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1018.925929] env[62965]: DEBUG nova.network.neutron [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 913c0536-5e39-4654-b083-2703a6aed328] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1018.946575] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390281, 'name': CreateVM_Task, 'duration_secs': 0.454115} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.946743] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1018.947503] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.947738] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.948120] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1018.948777] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75f7f54e-eb18-48dd-a728-81f8bb3ac45a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.956212] env[62965]: DEBUG oslo_vmware.api [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 1018.956212] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52f9954d-82a0-8a71-11b9-c022a067ea19" [ 1018.956212] env[62965]: _type = "Task" [ 1018.956212] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.969334] env[62965]: DEBUG oslo_vmware.api [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52f9954d-82a0-8a71-11b9-c022a067ea19, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.972702] env[62965]: DEBUG oslo_vmware.api [None req-f0be609a-e4c5-4ee3-9430-9d7a07622409 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390282, 'name': PowerOffVM_Task, 'duration_secs': 0.175496} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.973060] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0be609a-e4c5-4ee3-9430-9d7a07622409 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1018.973279] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-f0be609a-e4c5-4ee3-9430-9d7a07622409 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1018.973579] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-38c45280-1b1f-4097-9158-751af379bf0d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.993018] env[62965]: DEBUG nova.policy [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e7510e1ec5ee4253b67be988b6b5ab51', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5d2d71b59ede430f846520bf49249e4a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 1019.050965] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquiring lock "refresh_cache-9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.050965] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquired lock "refresh_cache-9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.050965] env[62965]: DEBUG nova.network.neutron [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1019.067161] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-f0be609a-e4c5-4ee3-9430-9d7a07622409 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1019.067161] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-f0be609a-e4c5-4ee3-9430-9d7a07622409 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Deleting contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1019.067161] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0be609a-e4c5-4ee3-9430-9d7a07622409 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Deleting the datastore file [datastore1] 0620c47c-2918-43dc-9e67-f32ca3dd20f1 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1019.067161] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4e1ca523-1ebb-4409-8783-278b422c3d83 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.073906] env[62965]: DEBUG oslo_vmware.api [None req-f0be609a-e4c5-4ee3-9430-9d7a07622409 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 1019.073906] env[62965]: value = "task-1390285" [ 1019.073906] env[62965]: _type = "Task" [ 1019.073906] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.083039] env[62965]: DEBUG oslo_vmware.api [None req-f0be609a-e4c5-4ee3-9430-9d7a07622409 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390285, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.118718] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4574bf1-bc0b-485a-a315-18ddc91f7f57 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.125289] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d29d4499-19cd-4f4e-86f8-ad16256b2a2b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.156031] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd772211-bc93-42ae-9f3a-32086faa651b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.163677] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c2cd61e-2d1a-499d-8ae6-8b709da272d0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.177150] env[62965]: DEBUG nova.compute.provider_tree [None req-7e818b1d-898d-46dc-9cff-8175b4ed1ce6 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1019.288543] env[62965]: DEBUG nova.network.neutron [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Successfully created port: 6579674f-a74a-48ac-b7b8-accba556f17f {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1019.333307] env[62965]: DEBUG oslo_vmware.api [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390283, 'name': PowerOffVM_Task, 'duration_secs': 0.22261} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.334080] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1019.334080] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1019.334080] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-03dc856b-41a8-42de-85bb-af87664567f3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.394875] env[62965]: DEBUG nova.network.neutron [-] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.431979] env[62965]: DEBUG nova.compute.manager [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1019.466022] env[62965]: DEBUG oslo_vmware.api [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52f9954d-82a0-8a71-11b9-c022a067ea19, 'name': SearchDatastore_Task, 'duration_secs': 0.01265} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.466421] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.466582] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1019.466896] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.467200] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.467534] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1019.467939] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-82c8ab02-3860-43d4-a8a7-faa3e7aeae54 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.479744] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1019.479744] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Deleting contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1019.479744] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Deleting the datastore file [datastore1] b59d6023-a047-4b32-bd18-1c7cc05f67d7 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1019.479744] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dd0af932-55e4-480f-958b-4e54bcbe0dad {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.479744] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1019.480056] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1019.480853] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-629d2309-db40-4e66-8fc0-8c7a5611bca7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.489688] env[62965]: DEBUG oslo_vmware.api [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 1019.489688] env[62965]: value = "task-1390287" [ 1019.489688] env[62965]: _type = "Task" [ 1019.489688] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.489688] env[62965]: DEBUG oslo_vmware.api [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 1019.489688] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52ff2a6e-4f5d-7872-7c9f-c03dbfae7401" [ 1019.489688] env[62965]: _type = "Task" [ 1019.489688] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.504851] env[62965]: DEBUG oslo_vmware.api [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390287, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.505132] env[62965]: DEBUG oslo_vmware.api [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52ff2a6e-4f5d-7872-7c9f-c03dbfae7401, 'name': SearchDatastore_Task, 'duration_secs': 0.008832} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.506309] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa5d59a2-151a-4788-ad23-0fcab38b91d0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.511155] env[62965]: DEBUG oslo_vmware.api [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 1019.511155] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52c8ce33-1e6f-aa37-6381-270098d0aa23" [ 1019.511155] env[62965]: _type = "Task" [ 1019.511155] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.518410] env[62965]: DEBUG oslo_vmware.api [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52c8ce33-1e6f-aa37-6381-270098d0aa23, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.529192] env[62965]: DEBUG nova.network.neutron [req-2f8ad426-89ab-4ddc-a0af-814bac57d4ea req-8a7cbccb-671c-49c9-a147-53a02b2a6037 service nova] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Updated VIF entry in instance network info cache for port 0870a7fe-8ea7-42dd-bbe4-558f7f5de034. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1019.529548] env[62965]: DEBUG nova.network.neutron [req-2f8ad426-89ab-4ddc-a0af-814bac57d4ea req-8a7cbccb-671c-49c9-a147-53a02b2a6037 service nova] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Updating instance_info_cache with network_info: [{"id": "0870a7fe-8ea7-42dd-bbe4-558f7f5de034", "address": "fa:16:3e:51:a1:8f", "network": {"id": "62052700-1781-4933-93c8-fa1aafdb0fd2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1931798228-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "978666785b0743bb843de96585f96fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0870a7fe-8e", "ovs_interfaceid": "0870a7fe-8ea7-42dd-bbe4-558f7f5de034", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.584163] env[62965]: DEBUG oslo_vmware.api [None req-f0be609a-e4c5-4ee3-9430-9d7a07622409 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390285, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.15307} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.584893] env[62965]: DEBUG nova.network.neutron [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1019.586697] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0be609a-e4c5-4ee3-9430-9d7a07622409 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1019.586932] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-f0be609a-e4c5-4ee3-9430-9d7a07622409 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Deleted contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1019.587135] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-f0be609a-e4c5-4ee3-9430-9d7a07622409 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1019.587316] env[62965]: INFO nova.compute.manager [None req-f0be609a-e4c5-4ee3-9430-9d7a07622409 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1019.587554] env[62965]: DEBUG oslo.service.loopingcall [None req-f0be609a-e4c5-4ee3-9430-9d7a07622409 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1019.587756] env[62965]: DEBUG nova.compute.manager [-] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1019.587851] env[62965]: DEBUG nova.network.neutron [-] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1019.679976] env[62965]: DEBUG nova.scheduler.client.report [None req-7e818b1d-898d-46dc-9cff-8175b4ed1ce6 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1019.804762] env[62965]: DEBUG nova.network.neutron [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Updating instance_info_cache with network_info: [{"id": "9f81f2ea-9520-4f9c-a2f4-e6c3b1ab1216", "address": "fa:16:3e:d5:77:3d", "network": {"id": "c4c8235c-333c-40e7-8744-110eaa0aa23c", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1256649418-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d2d71b59ede430f846520bf49249e4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f81f2ea-95", "ovs_interfaceid": "9f81f2ea-9520-4f9c-a2f4-e6c3b1ab1216", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.900021] env[62965]: INFO nova.compute.manager [-] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Took 1.33 seconds to deallocate network for instance. [ 1019.999019] env[62965]: DEBUG oslo_vmware.api [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390287, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.1227} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.999544] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1019.999737] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Deleted contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1019.999949] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1020.000149] env[62965]: INFO nova.compute.manager [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Took 1.20 seconds to destroy the instance on the hypervisor. [ 1020.000389] env[62965]: DEBUG oslo.service.loopingcall [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1020.000582] env[62965]: DEBUG nova.compute.manager [-] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1020.000678] env[62965]: DEBUG nova.network.neutron [-] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1020.020138] env[62965]: DEBUG oslo_vmware.api [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52c8ce33-1e6f-aa37-6381-270098d0aa23, 'name': SearchDatastore_Task, 'duration_secs': 0.007951} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.020403] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.020653] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 8ae1fde5-e94e-43ed-816a-9e22633dfdd1/8ae1fde5-e94e-43ed-816a-9e22633dfdd1.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1020.020934] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2b3cada8-6232-4a04-a05c-057366faf404 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.027241] env[62965]: DEBUG oslo_vmware.api [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 1020.027241] env[62965]: value = "task-1390288" [ 1020.027241] env[62965]: _type = "Task" [ 1020.027241] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.031743] env[62965]: DEBUG oslo_concurrency.lockutils [req-2f8ad426-89ab-4ddc-a0af-814bac57d4ea req-8a7cbccb-671c-49c9-a147-53a02b2a6037 service nova] Releasing lock "refresh_cache-8ae1fde5-e94e-43ed-816a-9e22633dfdd1" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.036612] env[62965]: DEBUG oslo_vmware.api [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390288, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.186664] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7e818b1d-898d-46dc-9cff-8175b4ed1ce6 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.766s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.191226] env[62965]: DEBUG oslo_concurrency.lockutils [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.109s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1020.191226] env[62965]: INFO nova.compute.claims [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1020.215598] env[62965]: INFO nova.scheduler.client.report [None req-7e818b1d-898d-46dc-9cff-8175b4ed1ce6 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Deleted allocations for instance a0e9cc1a-bc95-4a44-99c5-aabf85a373f4 [ 1020.309378] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Releasing lock "refresh_cache-9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.309706] env[62965]: DEBUG nova.compute.manager [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Instance network_info: |[{"id": "9f81f2ea-9520-4f9c-a2f4-e6c3b1ab1216", "address": "fa:16:3e:d5:77:3d", "network": {"id": "c4c8235c-333c-40e7-8744-110eaa0aa23c", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1256649418-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d2d71b59ede430f846520bf49249e4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f81f2ea-95", "ovs_interfaceid": "9f81f2ea-9520-4f9c-a2f4-e6c3b1ab1216", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1020.310208] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d5:77:3d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dc6d5964-1106-4345-a26d-185dabd4ff0f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9f81f2ea-9520-4f9c-a2f4-e6c3b1ab1216', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1020.318357] env[62965]: DEBUG oslo.service.loopingcall [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1020.318613] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1020.318871] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5c89983e-84b4-4673-a002-bbb2f470bd17 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.340283] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1020.340283] env[62965]: value = "task-1390289" [ 1020.340283] env[62965]: _type = "Task" [ 1020.340283] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.350920] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390289, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.406012] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e6919be8-1742-47e5-8e55-6cb95da9c828 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.409546] env[62965]: DEBUG nova.network.neutron [-] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.440626] env[62965]: DEBUG nova.compute.manager [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1020.472260] env[62965]: DEBUG nova.virt.hardware [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1020.472556] env[62965]: DEBUG nova.virt.hardware [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1020.472718] env[62965]: DEBUG nova.virt.hardware [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1020.472907] env[62965]: DEBUG nova.virt.hardware [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1020.473119] env[62965]: DEBUG nova.virt.hardware [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1020.473235] env[62965]: DEBUG nova.virt.hardware [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1020.473453] env[62965]: DEBUG nova.virt.hardware [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1020.473619] env[62965]: DEBUG nova.virt.hardware [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1020.473791] env[62965]: DEBUG nova.virt.hardware [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1020.473958] env[62965]: DEBUG nova.virt.hardware [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1020.474155] env[62965]: DEBUG nova.virt.hardware [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1020.475223] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ed53929-860d-4374-855c-e72cdfe03ab8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.487310] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e54d1ec5-3804-435b-8415-1b470ef80f4b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.497191] env[62965]: DEBUG nova.compute.manager [req-068c0887-0c06-4ac3-af17-7ad199dbf3b5 req-aea8c67b-124e-4f4e-96e6-6a086ead55c4 service nova] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Received event network-changed-9f81f2ea-9520-4f9c-a2f4-e6c3b1ab1216 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1020.497661] env[62965]: DEBUG nova.compute.manager [req-068c0887-0c06-4ac3-af17-7ad199dbf3b5 req-aea8c67b-124e-4f4e-96e6-6a086ead55c4 service nova] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Refreshing instance network info cache due to event network-changed-9f81f2ea-9520-4f9c-a2f4-e6c3b1ab1216. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 1020.497992] env[62965]: DEBUG oslo_concurrency.lockutils [req-068c0887-0c06-4ac3-af17-7ad199dbf3b5 req-aea8c67b-124e-4f4e-96e6-6a086ead55c4 service nova] Acquiring lock "refresh_cache-9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.498287] env[62965]: DEBUG oslo_concurrency.lockutils [req-068c0887-0c06-4ac3-af17-7ad199dbf3b5 req-aea8c67b-124e-4f4e-96e6-6a086ead55c4 service nova] Acquired lock "refresh_cache-9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.498650] env[62965]: DEBUG nova.network.neutron [req-068c0887-0c06-4ac3-af17-7ad199dbf3b5 req-aea8c67b-124e-4f4e-96e6-6a086ead55c4 service nova] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Refreshing network info cache for port 9f81f2ea-9520-4f9c-a2f4-e6c3b1ab1216 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1020.541820] env[62965]: DEBUG oslo_vmware.api [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390288, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.726333] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7e818b1d-898d-46dc-9cff-8175b4ed1ce6 tempest-ServersNegativeTestJSON-890816944 tempest-ServersNegativeTestJSON-890816944-project-member] Lock "a0e9cc1a-bc95-4a44-99c5-aabf85a373f4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.932s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.800139] env[62965]: DEBUG nova.network.neutron [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Successfully updated port: 6579674f-a74a-48ac-b7b8-accba556f17f {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1020.833121] env[62965]: DEBUG nova.compute.manager [req-3206b691-dd2b-4945-a79d-f0148d3ad26e req-061c1946-4d46-432b-ae07-b94428a43d11 service nova] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Received event network-vif-deleted-b178e278-58df-4607-ae3b-1d6c07cd22e7 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1020.833327] env[62965]: DEBUG nova.compute.manager [req-3206b691-dd2b-4945-a79d-f0148d3ad26e req-061c1946-4d46-432b-ae07-b94428a43d11 service nova] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Received event network-vif-deleted-d2d44211-6856-42ed-8f9f-3b1a58b31349 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1020.833499] env[62965]: INFO nova.compute.manager [req-3206b691-dd2b-4945-a79d-f0148d3ad26e req-061c1946-4d46-432b-ae07-b94428a43d11 service nova] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Neutron deleted interface d2d44211-6856-42ed-8f9f-3b1a58b31349; detaching it from the instance and deleting it from the info cache [ 1020.833674] env[62965]: DEBUG nova.network.neutron [req-3206b691-dd2b-4945-a79d-f0148d3ad26e req-061c1946-4d46-432b-ae07-b94428a43d11 service nova] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.850339] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390289, 'name': CreateVM_Task, 'duration_secs': 0.42128} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.850509] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1020.851710] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.851888] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.852232] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1020.852711] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88177f64-854e-4eb1-b83d-23b6b26eb648 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.857626] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 1020.857626] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52d3d00e-fc7f-bf49-0e11-e2bdc8e5ff6f" [ 1020.857626] env[62965]: _type = "Task" [ 1020.857626] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.868382] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52d3d00e-fc7f-bf49-0e11-e2bdc8e5ff6f, 'name': SearchDatastore_Task, 'duration_secs': 0.008703} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.868661] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.868890] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1020.871095] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1020.871095] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.871095] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1020.871095] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-727da6ad-2fe8-4142-819a-26fe1bf321f3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.877082] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1020.877267] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1020.879448] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b2b31f22-3df6-49fd-94be-0ca48f77b755 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.883224] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 1020.883224] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]522764c6-4e78-5d61-45af-3c844749457a" [ 1020.883224] env[62965]: _type = "Task" [ 1020.883224] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.890979] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]522764c6-4e78-5d61-45af-3c844749457a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.911880] env[62965]: INFO nova.compute.manager [-] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Took 1.32 seconds to deallocate network for instance. [ 1021.038264] env[62965]: DEBUG oslo_vmware.api [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390288, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.535854} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.038572] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 8ae1fde5-e94e-43ed-816a-9e22633dfdd1/8ae1fde5-e94e-43ed-816a-9e22633dfdd1.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1021.038847] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1021.039158] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-58c825c6-e7e5-422f-a8d4-7ca2e5c0430e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.044939] env[62965]: DEBUG oslo_vmware.api [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 1021.044939] env[62965]: value = "task-1390290" [ 1021.044939] env[62965]: _type = "Task" [ 1021.044939] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.052062] env[62965]: DEBUG oslo_vmware.api [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390290, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.184408] env[62965]: DEBUG nova.network.neutron [-] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.297477] env[62965]: DEBUG nova.network.neutron [req-068c0887-0c06-4ac3-af17-7ad199dbf3b5 req-aea8c67b-124e-4f4e-96e6-6a086ead55c4 service nova] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Updated VIF entry in instance network info cache for port 9f81f2ea-9520-4f9c-a2f4-e6c3b1ab1216. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1021.297994] env[62965]: DEBUG nova.network.neutron [req-068c0887-0c06-4ac3-af17-7ad199dbf3b5 req-aea8c67b-124e-4f4e-96e6-6a086ead55c4 service nova] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Updating instance_info_cache with network_info: [{"id": "9f81f2ea-9520-4f9c-a2f4-e6c3b1ab1216", "address": "fa:16:3e:d5:77:3d", "network": {"id": "c4c8235c-333c-40e7-8744-110eaa0aa23c", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1256649418-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d2d71b59ede430f846520bf49249e4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f81f2ea-95", "ovs_interfaceid": "9f81f2ea-9520-4f9c-a2f4-e6c3b1ab1216", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.346370] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquiring lock "refresh_cache-913c0536-5e39-4654-b083-2703a6aed328" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.346370] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquired lock "refresh_cache-913c0536-5e39-4654-b083-2703a6aed328" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.346370] env[62965]: DEBUG nova.network.neutron [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1021.346370] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7336412a-1d1c-45d4-8ca6-0e98196ba481 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.351383] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90b27b31-096b-4248-8b6a-24412d2b000d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.378335] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c152312f-fc5d-4c06-8afb-1633de17eaf2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.387966] env[62965]: DEBUG nova.compute.manager [req-3206b691-dd2b-4945-a79d-f0148d3ad26e req-061c1946-4d46-432b-ae07-b94428a43d11 service nova] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Detach interface failed, port_id=d2d44211-6856-42ed-8f9f-3b1a58b31349, reason: Instance b59d6023-a047-4b32-bd18-1c7cc05f67d7 could not be found. {{(pid=62965) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11294}} [ 1021.394421] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de335648-c357-41ed-8150-c021b3990c5b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.401365] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]522764c6-4e78-5d61-45af-3c844749457a, 'name': SearchDatastore_Task, 'duration_secs': 0.007435} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.402499] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bcae7cec-885c-419c-922f-326684e27e89 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.430149] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f0be609a-e4c5-4ee3-9430-9d7a07622409 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.431544] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e1fbf4d-859f-4783-9ec7-4be13f18e347 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.435280] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 1021.435280] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52bfb32c-aacd-1672-5fe3-2c8040ac414e" [ 1021.435280] env[62965]: _type = "Task" [ 1021.435280] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.442235] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ef336b9-6841-4543-aed3-79469223a150 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.449180] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52bfb32c-aacd-1672-5fe3-2c8040ac414e, 'name': SearchDatastore_Task, 'duration_secs': 0.00972} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.449724] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.450026] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b/9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1021.450272] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-37cf556d-5706-47f1-a56d-794b68e5135f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.459567] env[62965]: DEBUG nova.compute.provider_tree [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1021.466251] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 1021.466251] env[62965]: value = "task-1390291" [ 1021.466251] env[62965]: _type = "Task" [ 1021.466251] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.474252] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390291, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.555351] env[62965]: DEBUG oslo_vmware.api [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390290, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.0683} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.556031] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1021.556401] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a7ab2d2-d3aa-487a-970a-e4b15adc5f4f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.577947] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] 8ae1fde5-e94e-43ed-816a-9e22633dfdd1/8ae1fde5-e94e-43ed-816a-9e22633dfdd1.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1021.578232] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aee23071-0e9e-41f7-b596-4617d63a4bd7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.597547] env[62965]: DEBUG oslo_vmware.api [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 1021.597547] env[62965]: value = "task-1390292" [ 1021.597547] env[62965]: _type = "Task" [ 1021.597547] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.605365] env[62965]: DEBUG oslo_vmware.api [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390292, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.686819] env[62965]: INFO nova.compute.manager [-] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Took 1.69 seconds to deallocate network for instance. [ 1021.800465] env[62965]: DEBUG oslo_concurrency.lockutils [req-068c0887-0c06-4ac3-af17-7ad199dbf3b5 req-aea8c67b-124e-4f4e-96e6-6a086ead55c4 service nova] Releasing lock "refresh_cache-9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1021.800748] env[62965]: DEBUG nova.compute.manager [req-068c0887-0c06-4ac3-af17-7ad199dbf3b5 req-aea8c67b-124e-4f4e-96e6-6a086ead55c4 service nova] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Received event network-vif-deleted-ee7ad38e-9936-45bf-8e20-b20e5f7958bd {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1021.854405] env[62965]: DEBUG nova.network.neutron [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1021.962559] env[62965]: DEBUG nova.scheduler.client.report [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1021.982757] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390291, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.471151} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.983695] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b/9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1021.983967] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1021.984282] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bcd700b1-9294-4d82-b95e-7a04dca3fba0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.991822] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 1021.991822] env[62965]: value = "task-1390293" [ 1021.991822] env[62965]: _type = "Task" [ 1021.991822] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.000228] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390293, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.019557] env[62965]: DEBUG nova.network.neutron [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Updating instance_info_cache with network_info: [{"id": "6579674f-a74a-48ac-b7b8-accba556f17f", "address": "fa:16:3e:78:de:31", "network": {"id": "c4c8235c-333c-40e7-8744-110eaa0aa23c", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1256649418-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d2d71b59ede430f846520bf49249e4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6579674f-a7", "ovs_interfaceid": "6579674f-a74a-48ac-b7b8-accba556f17f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.108186] env[62965]: DEBUG oslo_vmware.api [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390292, 'name': ReconfigVM_Task, 'duration_secs': 0.48307} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.108472] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Reconfigured VM instance instance-00000061 to attach disk [datastore2] 8ae1fde5-e94e-43ed-816a-9e22633dfdd1/8ae1fde5-e94e-43ed-816a-9e22633dfdd1.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1022.109146] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2fe20a8e-03b3-4812-8543-f8c5db1699e2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.115457] env[62965]: DEBUG oslo_vmware.api [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 1022.115457] env[62965]: value = "task-1390294" [ 1022.115457] env[62965]: _type = "Task" [ 1022.115457] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.123596] env[62965]: DEBUG oslo_vmware.api [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390294, 'name': Rename_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.194442] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.467961] env[62965]: DEBUG oslo_concurrency.lockutils [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.279s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.469037] env[62965]: DEBUG nova.compute.manager [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1022.472640] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 6.641s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.472915] env[62965]: DEBUG nova.objects.instance [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62965) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1022.504900] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390293, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064312} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.505380] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1022.507398] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79c9d3af-d011-4d87-910d-13522002c201 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.531019] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Releasing lock "refresh_cache-913c0536-5e39-4654-b083-2703a6aed328" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.531019] env[62965]: DEBUG nova.compute.manager [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Instance network_info: |[{"id": "6579674f-a74a-48ac-b7b8-accba556f17f", "address": "fa:16:3e:78:de:31", "network": {"id": "c4c8235c-333c-40e7-8744-110eaa0aa23c", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1256649418-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d2d71b59ede430f846520bf49249e4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6579674f-a7", "ovs_interfaceid": "6579674f-a74a-48ac-b7b8-accba556f17f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1022.540406] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b/9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1022.542763] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:78:de:31', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dc6d5964-1106-4345-a26d-185dabd4ff0f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6579674f-a74a-48ac-b7b8-accba556f17f', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1022.550579] env[62965]: DEBUG oslo.service.loopingcall [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1022.550579] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f03d38bd-2ce2-4af7-8c22-7fb88dd6f789 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.573056] env[62965]: DEBUG nova.compute.manager [req-be2a9a12-2b7a-44a4-a73a-2894f4b225dc req-7addbb8d-72f4-45d2-8a1d-237306b40e8b service nova] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Received event network-vif-plugged-6579674f-a74a-48ac-b7b8-accba556f17f {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1022.573272] env[62965]: DEBUG oslo_concurrency.lockutils [req-be2a9a12-2b7a-44a4-a73a-2894f4b225dc req-7addbb8d-72f4-45d2-8a1d-237306b40e8b service nova] Acquiring lock "913c0536-5e39-4654-b083-2703a6aed328-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.573551] env[62965]: DEBUG oslo_concurrency.lockutils [req-be2a9a12-2b7a-44a4-a73a-2894f4b225dc req-7addbb8d-72f4-45d2-8a1d-237306b40e8b service nova] Lock "913c0536-5e39-4654-b083-2703a6aed328-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.573674] env[62965]: DEBUG oslo_concurrency.lockutils [req-be2a9a12-2b7a-44a4-a73a-2894f4b225dc req-7addbb8d-72f4-45d2-8a1d-237306b40e8b service nova] Lock "913c0536-5e39-4654-b083-2703a6aed328-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.576781] env[62965]: DEBUG nova.compute.manager [req-be2a9a12-2b7a-44a4-a73a-2894f4b225dc req-7addbb8d-72f4-45d2-8a1d-237306b40e8b service nova] [instance: 913c0536-5e39-4654-b083-2703a6aed328] No waiting events found dispatching network-vif-plugged-6579674f-a74a-48ac-b7b8-accba556f17f {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1022.576781] env[62965]: WARNING nova.compute.manager [req-be2a9a12-2b7a-44a4-a73a-2894f4b225dc req-7addbb8d-72f4-45d2-8a1d-237306b40e8b service nova] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Received unexpected event network-vif-plugged-6579674f-a74a-48ac-b7b8-accba556f17f for instance with vm_state building and task_state spawning. [ 1022.576781] env[62965]: DEBUG nova.compute.manager [req-be2a9a12-2b7a-44a4-a73a-2894f4b225dc req-7addbb8d-72f4-45d2-8a1d-237306b40e8b service nova] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Received event network-changed-6579674f-a74a-48ac-b7b8-accba556f17f {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1022.576781] env[62965]: DEBUG nova.compute.manager [req-be2a9a12-2b7a-44a4-a73a-2894f4b225dc req-7addbb8d-72f4-45d2-8a1d-237306b40e8b service nova] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Refreshing instance network info cache due to event network-changed-6579674f-a74a-48ac-b7b8-accba556f17f. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 1022.576781] env[62965]: DEBUG oslo_concurrency.lockutils [req-be2a9a12-2b7a-44a4-a73a-2894f4b225dc req-7addbb8d-72f4-45d2-8a1d-237306b40e8b service nova] Acquiring lock "refresh_cache-913c0536-5e39-4654-b083-2703a6aed328" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.576781] env[62965]: DEBUG oslo_concurrency.lockutils [req-be2a9a12-2b7a-44a4-a73a-2894f4b225dc req-7addbb8d-72f4-45d2-8a1d-237306b40e8b service nova] Acquired lock "refresh_cache-913c0536-5e39-4654-b083-2703a6aed328" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.576781] env[62965]: DEBUG nova.network.neutron [req-be2a9a12-2b7a-44a4-a73a-2894f4b225dc req-7addbb8d-72f4-45d2-8a1d-237306b40e8b service nova] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Refreshing network info cache for port 6579674f-a74a-48ac-b7b8-accba556f17f {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1022.576781] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1022.576781] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0408dc5a-2adf-49c2-9e3d-95e3b3e54fa6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.596206] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 1022.596206] env[62965]: value = "task-1390295" [ 1022.596206] env[62965]: _type = "Task" [ 1022.596206] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.600952] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1022.600952] env[62965]: value = "task-1390296" [ 1022.600952] env[62965]: _type = "Task" [ 1022.600952] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.607785] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390295, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.613114] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390296, 'name': CreateVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.625881] env[62965]: DEBUG oslo_vmware.api [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390294, 'name': Rename_Task, 'duration_secs': 0.13701} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.626175] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1022.626424] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5f08f874-4391-4ff1-8bd2-548aec20804a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.631782] env[62965]: DEBUG oslo_vmware.api [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 1022.631782] env[62965]: value = "task-1390297" [ 1022.631782] env[62965]: _type = "Task" [ 1022.631782] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.639708] env[62965]: DEBUG oslo_vmware.api [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390297, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.982807] env[62965]: DEBUG nova.compute.utils [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1022.986811] env[62965]: DEBUG nova.compute.manager [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1022.986974] env[62965]: DEBUG nova.network.neutron [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1023.036635] env[62965]: DEBUG nova.policy [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '34e15ad3c87445e5a217300a23223540', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8bfa5cfdc7d5496482d87675f2a122dd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 1023.109377] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390295, 'name': ReconfigVM_Task, 'duration_secs': 0.300526} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.110075] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Reconfigured VM instance instance-00000062 to attach disk [datastore2] 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b/9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1023.110828] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-09271e0c-2d44-4085-9e0c-e8d2adf637b4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.115853] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390296, 'name': CreateVM_Task, 'duration_secs': 0.331925} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.116649] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1023.117513] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1023.117971] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.118612] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1023.118612] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4aa7b8f2-aed2-42ff-8a7d-d9acd343987b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.122372] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 1023.122372] env[62965]: value = "task-1390298" [ 1023.122372] env[62965]: _type = "Task" [ 1023.122372] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.128831] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 1023.128831] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52dfa452-89fe-4d93-68e3-a53933ed6e59" [ 1023.128831] env[62965]: _type = "Task" [ 1023.128831] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.136020] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390298, 'name': Rename_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.146830] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52dfa452-89fe-4d93-68e3-a53933ed6e59, 'name': SearchDatastore_Task, 'duration_secs': 0.010949} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.150908] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.151117] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1023.151754] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1023.151939] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.152145] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1023.152447] env[62965]: DEBUG oslo_vmware.api [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390297, 'name': PowerOnVM_Task, 'duration_secs': 0.502193} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.152655] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c24e813b-865a-4715-a729-89814ed4eeaa {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.155085] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1023.155403] env[62965]: INFO nova.compute.manager [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Took 6.78 seconds to spawn the instance on the hypervisor. [ 1023.155651] env[62965]: DEBUG nova.compute.manager [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1023.156614] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a142baf2-b31a-4248-b4e6-2a9c270f07d3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.167434] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1023.167791] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1023.168948] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-878ad1fc-4a73-4eb5-9ffa-ec10bb466f26 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.174508] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 1023.174508] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]522ac74e-23de-cd67-0397-7aa966d33994" [ 1023.174508] env[62965]: _type = "Task" [ 1023.174508] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.184411] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]522ac74e-23de-cd67-0397-7aa966d33994, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.318323] env[62965]: DEBUG nova.network.neutron [req-be2a9a12-2b7a-44a4-a73a-2894f4b225dc req-7addbb8d-72f4-45d2-8a1d-237306b40e8b service nova] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Updated VIF entry in instance network info cache for port 6579674f-a74a-48ac-b7b8-accba556f17f. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1023.318699] env[62965]: DEBUG nova.network.neutron [req-be2a9a12-2b7a-44a4-a73a-2894f4b225dc req-7addbb8d-72f4-45d2-8a1d-237306b40e8b service nova] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Updating instance_info_cache with network_info: [{"id": "6579674f-a74a-48ac-b7b8-accba556f17f", "address": "fa:16:3e:78:de:31", "network": {"id": "c4c8235c-333c-40e7-8744-110eaa0aa23c", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1256649418-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5d2d71b59ede430f846520bf49249e4a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dc6d5964-1106-4345-a26d-185dabd4ff0f", "external-id": "nsx-vlan-transportzone-603", "segmentation_id": 603, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6579674f-a7", "ovs_interfaceid": "6579674f-a74a-48ac-b7b8-accba556f17f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.360797] env[62965]: DEBUG nova.network.neutron [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Successfully created port: 0ecbf02c-8b40-41fc-a928-d81ec8eeb2a0 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1023.488087] env[62965]: DEBUG nova.compute.manager [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1023.492207] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2371dea1-7dec-41e9-8f8d-3174e5312e5e tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.020s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.493154] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7a26068c-4b51-4f97-af71-622aad6bddf0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.648s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.493372] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7a26068c-4b51-4f97-af71-622aad6bddf0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.499495] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3dfc7706-1742-4cc3-8097-297bcd700574 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.751s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.499732] env[62965]: DEBUG nova.objects.instance [None req-3dfc7706-1742-4cc3-8097-297bcd700574 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lazy-loading 'resources' on Instance uuid a63b301d-8d52-4fec-951f-a15cda9449ac {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1023.537399] env[62965]: INFO nova.scheduler.client.report [None req-7a26068c-4b51-4f97-af71-622aad6bddf0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Deleted allocations for instance 464a3a31-06bf-4662-a5b1-d185d9a64f0b [ 1023.633170] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390298, 'name': Rename_Task, 'duration_secs': 0.153124} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.633527] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1023.635072] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ad32152c-c164-4c68-8f79-f58a50a1964f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.642199] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 1023.642199] env[62965]: value = "task-1390299" [ 1023.642199] env[62965]: _type = "Task" [ 1023.642199] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.652708] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390299, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.677855] env[62965]: INFO nova.compute.manager [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Took 15.39 seconds to build instance. [ 1023.689204] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]522ac74e-23de-cd67-0397-7aa966d33994, 'name': SearchDatastore_Task, 'duration_secs': 0.008648} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.690334] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cced0c6a-0dfb-400b-a0ec-5dcae7f2bf1c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.696218] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 1023.696218] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]526f99b0-8afe-a8be-736d-2894c6e897f9" [ 1023.696218] env[62965]: _type = "Task" [ 1023.696218] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.704960] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]526f99b0-8afe-a8be-736d-2894c6e897f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.821550] env[62965]: DEBUG oslo_concurrency.lockutils [req-be2a9a12-2b7a-44a4-a73a-2894f4b225dc req-7addbb8d-72f4-45d2-8a1d-237306b40e8b service nova] Releasing lock "refresh_cache-913c0536-5e39-4654-b083-2703a6aed328" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.047525] env[62965]: DEBUG oslo_concurrency.lockutils [None req-7a26068c-4b51-4f97-af71-622aad6bddf0 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "464a3a31-06bf-4662-a5b1-d185d9a64f0b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.599s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.139978] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d70ada5c-717a-4698-b23e-876115cfb64a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.144729] env[62965]: DEBUG oslo_concurrency.lockutils [None req-673d91ef-6bee-4ba7-ba0d-80a3c8cafc7f tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "8ae1fde5-e94e-43ed-816a-9e22633dfdd1" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.150092] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0810dcfd-44d3-4041-a20a-acdc03aa0a2a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.157703] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390299, 'name': PowerOnVM_Task, 'duration_secs': 0.505033} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.157984] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1024.158219] env[62965]: INFO nova.compute.manager [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Took 5.99 seconds to spawn the instance on the hypervisor. [ 1024.158456] env[62965]: DEBUG nova.compute.manager [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1024.182893] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-396d45eb-40e9-45e8-8ad3-3973a2579265 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.186349] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4f7280d8-92a0-4ee2-afbb-0fbe20dc3621 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "8ae1fde5-e94e-43ed-816a-9e22633dfdd1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.908s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.186919] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30e6c672-0d44-4134-88ef-77a483d5c8c3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.189266] env[62965]: DEBUG oslo_concurrency.lockutils [None req-673d91ef-6bee-4ba7-ba0d-80a3c8cafc7f tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "8ae1fde5-e94e-43ed-816a-9e22633dfdd1" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.045s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.189466] env[62965]: DEBUG nova.compute.manager [None req-673d91ef-6bee-4ba7-ba0d-80a3c8cafc7f tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1024.190625] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4436ba97-3bc1-431f-91ce-942e2db7517d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.204770] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffdaff28-9e5b-4921-907c-dd62b2791cef {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.210575] env[62965]: DEBUG nova.compute.manager [None req-673d91ef-6bee-4ba7-ba0d-80a3c8cafc7f tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62965) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 1024.211084] env[62965]: DEBUG nova.objects.instance [None req-673d91ef-6bee-4ba7-ba0d-80a3c8cafc7f tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lazy-loading 'flavor' on Instance uuid 8ae1fde5-e94e-43ed-816a-9e22633dfdd1 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1024.218510] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]526f99b0-8afe-a8be-736d-2894c6e897f9, 'name': SearchDatastore_Task, 'duration_secs': 0.008945} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.229370] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.229665] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 913c0536-5e39-4654-b083-2703a6aed328/913c0536-5e39-4654-b083-2703a6aed328.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1024.230157] env[62965]: DEBUG nova.compute.provider_tree [None req-3dfc7706-1742-4cc3-8097-297bcd700574 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1024.234636] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cad310de-5e61-4572-aa0b-77e1da933c51 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.241430] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 1024.241430] env[62965]: value = "task-1390300" [ 1024.241430] env[62965]: _type = "Task" [ 1024.241430] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.249306] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390300, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.497927] env[62965]: DEBUG nova.compute.manager [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1024.525777] env[62965]: DEBUG nova.virt.hardware [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1024.526051] env[62965]: DEBUG nova.virt.hardware [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1024.526226] env[62965]: DEBUG nova.virt.hardware [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1024.526415] env[62965]: DEBUG nova.virt.hardware [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1024.526560] env[62965]: DEBUG nova.virt.hardware [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1024.526701] env[62965]: DEBUG nova.virt.hardware [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1024.526913] env[62965]: DEBUG nova.virt.hardware [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1024.527085] env[62965]: DEBUG nova.virt.hardware [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1024.527267] env[62965]: DEBUG nova.virt.hardware [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1024.527433] env[62965]: DEBUG nova.virt.hardware [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1024.527603] env[62965]: DEBUG nova.virt.hardware [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1024.528518] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b06837e-147b-497d-89bd-16e37a5ffbb8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.537185] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf5876cb-3d59-4886-9520-c62c6587d685 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.711735] env[62965]: INFO nova.compute.manager [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Took 12.59 seconds to build instance. [ 1024.736847] env[62965]: DEBUG nova.scheduler.client.report [None req-3dfc7706-1742-4cc3-8097-297bcd700574 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1024.751337] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390300, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.464717} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.751588] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 913c0536-5e39-4654-b083-2703a6aed328/913c0536-5e39-4654-b083-2703a6aed328.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1024.751800] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1024.752060] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2aefc23a-3394-4f71-a18e-45c69cbc4680 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.758924] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 1024.758924] env[62965]: value = "task-1390301" [ 1024.758924] env[62965]: _type = "Task" [ 1024.758924] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.766338] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390301, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.000341] env[62965]: DEBUG nova.compute.manager [req-085f426a-d30e-4d7e-936d-e7f074b742ea req-b3c093ce-bec7-441f-b3ff-38792ee6d24c service nova] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Received event network-vif-plugged-0ecbf02c-8b40-41fc-a928-d81ec8eeb2a0 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1025.000652] env[62965]: DEBUG oslo_concurrency.lockutils [req-085f426a-d30e-4d7e-936d-e7f074b742ea req-b3c093ce-bec7-441f-b3ff-38792ee6d24c service nova] Acquiring lock "df03e327-959c-42af-85cd-a7af2ec4d0f7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.000782] env[62965]: DEBUG oslo_concurrency.lockutils [req-085f426a-d30e-4d7e-936d-e7f074b742ea req-b3c093ce-bec7-441f-b3ff-38792ee6d24c service nova] Lock "df03e327-959c-42af-85cd-a7af2ec4d0f7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.001028] env[62965]: DEBUG oslo_concurrency.lockutils [req-085f426a-d30e-4d7e-936d-e7f074b742ea req-b3c093ce-bec7-441f-b3ff-38792ee6d24c service nova] Lock "df03e327-959c-42af-85cd-a7af2ec4d0f7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.001220] env[62965]: DEBUG nova.compute.manager [req-085f426a-d30e-4d7e-936d-e7f074b742ea req-b3c093ce-bec7-441f-b3ff-38792ee6d24c service nova] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] No waiting events found dispatching network-vif-plugged-0ecbf02c-8b40-41fc-a928-d81ec8eeb2a0 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1025.001368] env[62965]: WARNING nova.compute.manager [req-085f426a-d30e-4d7e-936d-e7f074b742ea req-b3c093ce-bec7-441f-b3ff-38792ee6d24c service nova] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Received unexpected event network-vif-plugged-0ecbf02c-8b40-41fc-a928-d81ec8eeb2a0 for instance with vm_state building and task_state spawning. [ 1025.088551] env[62965]: DEBUG nova.network.neutron [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Successfully updated port: 0ecbf02c-8b40-41fc-a928-d81ec8eeb2a0 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1025.214305] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.651s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.220051] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-673d91ef-6bee-4ba7-ba0d-80a3c8cafc7f tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1025.220335] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eab7e56e-6afe-4f65-84f2-ae539eebf541 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.227639] env[62965]: DEBUG oslo_vmware.api [None req-673d91ef-6bee-4ba7-ba0d-80a3c8cafc7f tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 1025.227639] env[62965]: value = "task-1390302" [ 1025.227639] env[62965]: _type = "Task" [ 1025.227639] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.236865] env[62965]: DEBUG oslo_vmware.api [None req-673d91ef-6bee-4ba7-ba0d-80a3c8cafc7f tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390302, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.240709] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3dfc7706-1742-4cc3-8097-297bcd700574 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.741s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.242963] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e6919be8-1742-47e5-8e55-6cb95da9c828 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.837s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.243229] env[62965]: DEBUG nova.objects.instance [None req-e6919be8-1742-47e5-8e55-6cb95da9c828 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lazy-loading 'resources' on Instance uuid 8efad776-6cbe-49ab-83dc-a410d8417505 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1025.263668] env[62965]: INFO nova.scheduler.client.report [None req-3dfc7706-1742-4cc3-8097-297bcd700574 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Deleted allocations for instance a63b301d-8d52-4fec-951f-a15cda9449ac [ 1025.271412] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390301, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067457} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.271649] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1025.272472] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18f8dd61-97ed-432b-9454-8f7978ff8d0d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.297159] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] 913c0536-5e39-4654-b083-2703a6aed328/913c0536-5e39-4654-b083-2703a6aed328.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1025.297484] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d934b24c-f6df-4757-8118-2cd4916a75f0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.318705] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 1025.318705] env[62965]: value = "task-1390303" [ 1025.318705] env[62965]: _type = "Task" [ 1025.318705] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.327311] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390303, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.591231] env[62965]: DEBUG oslo_concurrency.lockutils [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "refresh_cache-df03e327-959c-42af-85cd-a7af2ec4d0f7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.591389] env[62965]: DEBUG oslo_concurrency.lockutils [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquired lock "refresh_cache-df03e327-959c-42af-85cd-a7af2ec4d0f7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.591538] env[62965]: DEBUG nova.network.neutron [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1025.738140] env[62965]: DEBUG oslo_vmware.api [None req-673d91ef-6bee-4ba7-ba0d-80a3c8cafc7f tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390302, 'name': PowerOffVM_Task, 'duration_secs': 0.197896} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.738433] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-673d91ef-6bee-4ba7-ba0d-80a3c8cafc7f tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1025.738682] env[62965]: DEBUG nova.compute.manager [None req-673d91ef-6bee-4ba7-ba0d-80a3c8cafc7f tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1025.739462] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16a26cb6-1194-48a9-9f6c-c72838cb40c9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.774885] env[62965]: DEBUG oslo_concurrency.lockutils [None req-3dfc7706-1742-4cc3-8097-297bcd700574 tempest-AttachVolumeNegativeTest-1942544788 tempest-AttachVolumeNegativeTest-1942544788-project-member] Lock "a63b301d-8d52-4fec-951f-a15cda9449ac" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.807s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.831060] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390303, 'name': ReconfigVM_Task, 'duration_secs': 0.26595} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.831379] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Reconfigured VM instance instance-00000063 to attach disk [datastore2] 913c0536-5e39-4654-b083-2703a6aed328/913c0536-5e39-4654-b083-2703a6aed328.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1025.832022] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e3261d37-f72e-4ac8-8696-e2a394f22605 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.838725] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 1025.838725] env[62965]: value = "task-1390304" [ 1025.838725] env[62965]: _type = "Task" [ 1025.838725] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.846384] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390304, 'name': Rename_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.882408] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e053b071-2943-4dbc-8e6d-19623e828031 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.889941] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ce9d04c-5873-4661-8649-4d706b4d73ef {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.921452] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7fe6a1f-b8b0-4247-909d-338d67314896 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.929442] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a28814b7-b68e-4792-9bbd-1946943258f7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.943260] env[62965]: DEBUG nova.compute.provider_tree [None req-e6919be8-1742-47e5-8e55-6cb95da9c828 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1026.134938] env[62965]: DEBUG nova.network.neutron [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1026.252980] env[62965]: DEBUG oslo_concurrency.lockutils [None req-673d91ef-6bee-4ba7-ba0d-80a3c8cafc7f tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "8ae1fde5-e94e-43ed-816a-9e22633dfdd1" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.064s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.320642] env[62965]: DEBUG nova.network.neutron [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Updating instance_info_cache with network_info: [{"id": "0ecbf02c-8b40-41fc-a928-d81ec8eeb2a0", "address": "fa:16:3e:5d:51:81", "network": {"id": "a5246cc1-6001-4c20-9724-fd36a0437582", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-644702631-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bfa5cfdc7d5496482d87675f2a122dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0ecbf02c-8b", "ovs_interfaceid": "0ecbf02c-8b40-41fc-a928-d81ec8eeb2a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.348114] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390304, 'name': Rename_Task, 'duration_secs': 0.135656} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.348406] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1026.348668] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-00bb6a72-a55b-4fb2-aa27-c3f74bc754bf {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.354870] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 1026.354870] env[62965]: value = "task-1390306" [ 1026.354870] env[62965]: _type = "Task" [ 1026.354870] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.362636] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390306, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.479351] env[62965]: DEBUG nova.scheduler.client.report [None req-e6919be8-1742-47e5-8e55-6cb95da9c828 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Updated inventory for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with generation 124 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 1026.479656] env[62965]: DEBUG nova.compute.provider_tree [None req-e6919be8-1742-47e5-8e55-6cb95da9c828 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Updating resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 generation from 124 to 125 during operation: update_inventory {{(pid=62965) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1026.479859] env[62965]: DEBUG nova.compute.provider_tree [None req-e6919be8-1742-47e5-8e55-6cb95da9c828 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1026.581565] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8332c061-e6d5-494e-ae5d-52f86cc0afc3 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "8ae1fde5-e94e-43ed-816a-9e22633dfdd1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.581949] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8332c061-e6d5-494e-ae5d-52f86cc0afc3 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "8ae1fde5-e94e-43ed-816a-9e22633dfdd1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.582126] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8332c061-e6d5-494e-ae5d-52f86cc0afc3 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "8ae1fde5-e94e-43ed-816a-9e22633dfdd1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.582326] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8332c061-e6d5-494e-ae5d-52f86cc0afc3 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "8ae1fde5-e94e-43ed-816a-9e22633dfdd1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.582503] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8332c061-e6d5-494e-ae5d-52f86cc0afc3 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "8ae1fde5-e94e-43ed-816a-9e22633dfdd1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.585296] env[62965]: INFO nova.compute.manager [None req-8332c061-e6d5-494e-ae5d-52f86cc0afc3 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Terminating instance [ 1026.823696] env[62965]: DEBUG oslo_concurrency.lockutils [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Releasing lock "refresh_cache-df03e327-959c-42af-85cd-a7af2ec4d0f7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.824188] env[62965]: DEBUG nova.compute.manager [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Instance network_info: |[{"id": "0ecbf02c-8b40-41fc-a928-d81ec8eeb2a0", "address": "fa:16:3e:5d:51:81", "network": {"id": "a5246cc1-6001-4c20-9724-fd36a0437582", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-644702631-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bfa5cfdc7d5496482d87675f2a122dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0ecbf02c-8b", "ovs_interfaceid": "0ecbf02c-8b40-41fc-a928-d81ec8eeb2a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1026.824639] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5d:51:81', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '61b8f0db-488e-42d7-bf6c-6c1665cd5616', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0ecbf02c-8b40-41fc-a928-d81ec8eeb2a0', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1026.832996] env[62965]: DEBUG oslo.service.loopingcall [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1026.833283] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1026.833530] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-20b9d5f0-4448-486e-b5ef-8bb53d265c14 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.854234] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1026.854234] env[62965]: value = "task-1390307" [ 1026.854234] env[62965]: _type = "Task" [ 1026.854234] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.867693] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390307, 'name': CreateVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.871290] env[62965]: DEBUG oslo_vmware.api [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390306, 'name': PowerOnVM_Task, 'duration_secs': 0.466971} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.871606] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1026.871833] env[62965]: INFO nova.compute.manager [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Took 6.43 seconds to spawn the instance on the hypervisor. [ 1026.872072] env[62965]: DEBUG nova.compute.manager [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1026.872920] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0972812e-41fc-4d5e-8cab-5431de3666c4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.984879] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e6919be8-1742-47e5-8e55-6cb95da9c828 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.742s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.987254] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f0be609a-e4c5-4ee3-9430-9d7a07622409 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.557s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.987507] env[62965]: DEBUG nova.objects.instance [None req-f0be609a-e4c5-4ee3-9430-9d7a07622409 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lazy-loading 'resources' on Instance uuid 0620c47c-2918-43dc-9e67-f32ca3dd20f1 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1027.011556] env[62965]: INFO nova.scheduler.client.report [None req-e6919be8-1742-47e5-8e55-6cb95da9c828 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Deleted allocations for instance 8efad776-6cbe-49ab-83dc-a410d8417505 [ 1027.030016] env[62965]: DEBUG nova.compute.manager [req-03a938ab-02e8-4d08-919e-7b0841c1fc69 req-421c2182-adca-4388-8751-54ef542a4ef3 service nova] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Received event network-changed-0ecbf02c-8b40-41fc-a928-d81ec8eeb2a0 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1027.030317] env[62965]: DEBUG nova.compute.manager [req-03a938ab-02e8-4d08-919e-7b0841c1fc69 req-421c2182-adca-4388-8751-54ef542a4ef3 service nova] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Refreshing instance network info cache due to event network-changed-0ecbf02c-8b40-41fc-a928-d81ec8eeb2a0. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 1027.030473] env[62965]: DEBUG oslo_concurrency.lockutils [req-03a938ab-02e8-4d08-919e-7b0841c1fc69 req-421c2182-adca-4388-8751-54ef542a4ef3 service nova] Acquiring lock "refresh_cache-df03e327-959c-42af-85cd-a7af2ec4d0f7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.030624] env[62965]: DEBUG oslo_concurrency.lockutils [req-03a938ab-02e8-4d08-919e-7b0841c1fc69 req-421c2182-adca-4388-8751-54ef542a4ef3 service nova] Acquired lock "refresh_cache-df03e327-959c-42af-85cd-a7af2ec4d0f7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.030788] env[62965]: DEBUG nova.network.neutron [req-03a938ab-02e8-4d08-919e-7b0841c1fc69 req-421c2182-adca-4388-8751-54ef542a4ef3 service nova] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Refreshing network info cache for port 0ecbf02c-8b40-41fc-a928-d81ec8eeb2a0 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1027.091732] env[62965]: DEBUG nova.compute.manager [None req-8332c061-e6d5-494e-ae5d-52f86cc0afc3 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1027.091962] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-8332c061-e6d5-494e-ae5d-52f86cc0afc3 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1027.093211] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc4012e1-5fc8-4d51-8784-24e387eaf20d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.101200] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-8332c061-e6d5-494e-ae5d-52f86cc0afc3 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1027.101458] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2627c4e5-3cfe-4f6b-8aa6-d91e0595efb9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.164438] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-8332c061-e6d5-494e-ae5d-52f86cc0afc3 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1027.164664] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-8332c061-e6d5-494e-ae5d-52f86cc0afc3 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Deleting contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1027.164834] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-8332c061-e6d5-494e-ae5d-52f86cc0afc3 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Deleting the datastore file [datastore2] 8ae1fde5-e94e-43ed-816a-9e22633dfdd1 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1027.165137] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-daf5b560-f074-4336-b61b-594a6c4b3bb5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.171275] env[62965]: DEBUG oslo_vmware.api [None req-8332c061-e6d5-494e-ae5d-52f86cc0afc3 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 1027.171275] env[62965]: value = "task-1390309" [ 1027.171275] env[62965]: _type = "Task" [ 1027.171275] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.179802] env[62965]: DEBUG oslo_vmware.api [None req-8332c061-e6d5-494e-ae5d-52f86cc0afc3 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390309, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.367574] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390307, 'name': CreateVM_Task, 'duration_secs': 0.511611} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.367816] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1027.368736] env[62965]: DEBUG oslo_concurrency.lockutils [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.368984] env[62965]: DEBUG oslo_concurrency.lockutils [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.369470] env[62965]: DEBUG oslo_concurrency.lockutils [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1027.369873] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d5b92f6-fe93-49dd-8a2c-4b04e087f7cd {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.375045] env[62965]: DEBUG oslo_vmware.api [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 1027.375045] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52cdc98a-eaa6-7e11-74ad-b1bcad06353d" [ 1027.375045] env[62965]: _type = "Task" [ 1027.375045] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.390875] env[62965]: DEBUG oslo_vmware.api [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52cdc98a-eaa6-7e11-74ad-b1bcad06353d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.394903] env[62965]: INFO nova.compute.manager [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Took 15.28 seconds to build instance. [ 1027.520480] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e6919be8-1742-47e5-8e55-6cb95da9c828 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "8efad776-6cbe-49ab-83dc-a410d8417505" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.653s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.604958] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b0bbaa0-0c35-4787-b81b-12e66db71d97 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.612369] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e00dcaf-3c53-4458-a876-94f2a530ce65 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.645362] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-865f8821-7d90-4976-8613-a0138355c8de {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.653208] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36d1b978-8725-4d9a-b825-649973ecd150 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.667205] env[62965]: DEBUG nova.compute.provider_tree [None req-f0be609a-e4c5-4ee3-9430-9d7a07622409 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1027.680214] env[62965]: DEBUG oslo_vmware.api [None req-8332c061-e6d5-494e-ae5d-52f86cc0afc3 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390309, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.274812} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.680463] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-8332c061-e6d5-494e-ae5d-52f86cc0afc3 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1027.680651] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-8332c061-e6d5-494e-ae5d-52f86cc0afc3 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Deleted contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1027.680831] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-8332c061-e6d5-494e-ae5d-52f86cc0afc3 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1027.681048] env[62965]: INFO nova.compute.manager [None req-8332c061-e6d5-494e-ae5d-52f86cc0afc3 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Took 0.59 seconds to destroy the instance on the hypervisor. [ 1027.681301] env[62965]: DEBUG oslo.service.loopingcall [None req-8332c061-e6d5-494e-ae5d-52f86cc0afc3 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1027.681496] env[62965]: DEBUG nova.compute.manager [-] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1027.681590] env[62965]: DEBUG nova.network.neutron [-] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1027.885013] env[62965]: DEBUG oslo_vmware.api [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52cdc98a-eaa6-7e11-74ad-b1bcad06353d, 'name': SearchDatastore_Task, 'duration_secs': 0.018243} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.885362] env[62965]: DEBUG oslo_concurrency.lockutils [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.885595] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1027.886025] env[62965]: DEBUG oslo_concurrency.lockutils [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.886025] env[62965]: DEBUG oslo_concurrency.lockutils [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.886156] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1027.886425] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-04deb472-2f0e-44f8-aa78-991da826c319 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.896302] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1027.896442] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1027.900785] env[62965]: DEBUG oslo_concurrency.lockutils [None req-c7dc3354-805f-45b2-9eae-907e935a45f0 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "913c0536-5e39-4654-b083-2703a6aed328" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.297s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.900785] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9e4a924-d081-44cc-889d-54d95f86819b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.905147] env[62965]: DEBUG oslo_vmware.api [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 1027.905147] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]521d2eb1-edc8-4bcd-13c1-3839dbe1cb60" [ 1027.905147] env[62965]: _type = "Task" [ 1027.905147] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.913661] env[62965]: DEBUG oslo_vmware.api [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]521d2eb1-edc8-4bcd-13c1-3839dbe1cb60, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.021372] env[62965]: DEBUG nova.compute.manager [req-9bfbd9d4-9e02-4d8c-8f86-a880c2641380 req-cecb2236-26d3-495e-b562-798397beef8a service nova] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Received event network-vif-deleted-0870a7fe-8ea7-42dd-bbe4-558f7f5de034 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1028.021536] env[62965]: INFO nova.compute.manager [req-9bfbd9d4-9e02-4d8c-8f86-a880c2641380 req-cecb2236-26d3-495e-b562-798397beef8a service nova] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Neutron deleted interface 0870a7fe-8ea7-42dd-bbe4-558f7f5de034; detaching it from the instance and deleting it from the info cache [ 1028.021749] env[62965]: DEBUG nova.network.neutron [req-9bfbd9d4-9e02-4d8c-8f86-a880c2641380 req-cecb2236-26d3-495e-b562-798397beef8a service nova] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.106137] env[62965]: DEBUG nova.network.neutron [req-03a938ab-02e8-4d08-919e-7b0841c1fc69 req-421c2182-adca-4388-8751-54ef542a4ef3 service nova] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Updated VIF entry in instance network info cache for port 0ecbf02c-8b40-41fc-a928-d81ec8eeb2a0. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1028.106382] env[62965]: DEBUG nova.network.neutron [req-03a938ab-02e8-4d08-919e-7b0841c1fc69 req-421c2182-adca-4388-8751-54ef542a4ef3 service nova] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Updating instance_info_cache with network_info: [{"id": "0ecbf02c-8b40-41fc-a928-d81ec8eeb2a0", "address": "fa:16:3e:5d:51:81", "network": {"id": "a5246cc1-6001-4c20-9724-fd36a0437582", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-644702631-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8bfa5cfdc7d5496482d87675f2a122dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "61b8f0db-488e-42d7-bf6c-6c1665cd5616", "external-id": "nsx-vlan-transportzone-655", "segmentation_id": 655, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0ecbf02c-8b", "ovs_interfaceid": "0ecbf02c-8b40-41fc-a928-d81ec8eeb2a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.175391] env[62965]: DEBUG nova.scheduler.client.report [None req-f0be609a-e4c5-4ee3-9430-9d7a07622409 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1028.415482] env[62965]: DEBUG oslo_vmware.api [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]521d2eb1-edc8-4bcd-13c1-3839dbe1cb60, 'name': SearchDatastore_Task, 'duration_secs': 0.02014} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.416283] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8e5fe61-7c46-4339-8844-0d0b129e0588 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.419991] env[62965]: DEBUG nova.network.neutron [-] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.422309] env[62965]: DEBUG oslo_vmware.api [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 1028.422309] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52fec821-ddce-651b-9cb1-04a4e0128ae1" [ 1028.422309] env[62965]: _type = "Task" [ 1028.422309] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.431489] env[62965]: DEBUG oslo_vmware.api [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52fec821-ddce-651b-9cb1-04a4e0128ae1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.525155] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e0ecda78-92ca-4b3e-a78c-8b9d85271036 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.534617] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5507e696-dd7c-4479-9311-b9685d6b874c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.562181] env[62965]: DEBUG nova.compute.manager [req-9bfbd9d4-9e02-4d8c-8f86-a880c2641380 req-cecb2236-26d3-495e-b562-798397beef8a service nova] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Detach interface failed, port_id=0870a7fe-8ea7-42dd-bbe4-558f7f5de034, reason: Instance 8ae1fde5-e94e-43ed-816a-9e22633dfdd1 could not be found. {{(pid=62965) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11294}} [ 1028.609197] env[62965]: DEBUG oslo_concurrency.lockutils [req-03a938ab-02e8-4d08-919e-7b0841c1fc69 req-421c2182-adca-4388-8751-54ef542a4ef3 service nova] Releasing lock "refresh_cache-df03e327-959c-42af-85cd-a7af2ec4d0f7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.677588] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f0be609a-e4c5-4ee3-9430-9d7a07622409 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.690s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.679880] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.486s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.680132] env[62965]: DEBUG nova.objects.instance [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lazy-loading 'resources' on Instance uuid b59d6023-a047-4b32-bd18-1c7cc05f67d7 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1028.693223] env[62965]: INFO nova.scheduler.client.report [None req-f0be609a-e4c5-4ee3-9430-9d7a07622409 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Deleted allocations for instance 0620c47c-2918-43dc-9e67-f32ca3dd20f1 [ 1028.737074] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d9fac0bd-f0d3-4ee4-9139-647b506fccc5 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquiring lock "9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.737316] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d9fac0bd-f0d3-4ee4-9139-647b506fccc5 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.737528] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d9fac0bd-f0d3-4ee4-9139-647b506fccc5 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquiring lock "9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.737711] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d9fac0bd-f0d3-4ee4-9139-647b506fccc5 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.737883] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d9fac0bd-f0d3-4ee4-9139-647b506fccc5 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.740685] env[62965]: INFO nova.compute.manager [None req-d9fac0bd-f0d3-4ee4-9139-647b506fccc5 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Terminating instance [ 1028.799786] env[62965]: DEBUG oslo_concurrency.lockutils [None req-667ab7c8-8108-40fe-ae71-ea26a63829ce tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquiring lock "913c0536-5e39-4654-b083-2703a6aed328" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.800151] env[62965]: DEBUG oslo_concurrency.lockutils [None req-667ab7c8-8108-40fe-ae71-ea26a63829ce tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "913c0536-5e39-4654-b083-2703a6aed328" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.800395] env[62965]: DEBUG oslo_concurrency.lockutils [None req-667ab7c8-8108-40fe-ae71-ea26a63829ce tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquiring lock "913c0536-5e39-4654-b083-2703a6aed328-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.800603] env[62965]: DEBUG oslo_concurrency.lockutils [None req-667ab7c8-8108-40fe-ae71-ea26a63829ce tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "913c0536-5e39-4654-b083-2703a6aed328-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.800772] env[62965]: DEBUG oslo_concurrency.lockutils [None req-667ab7c8-8108-40fe-ae71-ea26a63829ce tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "913c0536-5e39-4654-b083-2703a6aed328-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.802860] env[62965]: INFO nova.compute.manager [None req-667ab7c8-8108-40fe-ae71-ea26a63829ce tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Terminating instance [ 1028.923805] env[62965]: INFO nova.compute.manager [-] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Took 1.24 seconds to deallocate network for instance. [ 1028.937619] env[62965]: DEBUG oslo_vmware.api [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52fec821-ddce-651b-9cb1-04a4e0128ae1, 'name': SearchDatastore_Task, 'duration_secs': 0.009419} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.937856] env[62965]: DEBUG oslo_concurrency.lockutils [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.938125] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] df03e327-959c-42af-85cd-a7af2ec4d0f7/df03e327-959c-42af-85cd-a7af2ec4d0f7.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1028.938380] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ee7ce41d-d26d-40b3-8019-d0110ac4353f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.945249] env[62965]: DEBUG oslo_vmware.api [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 1028.945249] env[62965]: value = "task-1390310" [ 1028.945249] env[62965]: _type = "Task" [ 1028.945249] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.960725] env[62965]: DEBUG oslo_vmware.api [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390310, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.203734] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f0be609a-e4c5-4ee3-9430-9d7a07622409 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "0620c47c-2918-43dc-9e67-f32ca3dd20f1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.282s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.246932] env[62965]: DEBUG nova.compute.manager [None req-d9fac0bd-f0d3-4ee4-9139-647b506fccc5 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1029.247096] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-d9fac0bd-f0d3-4ee4-9139-647b506fccc5 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1029.248196] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0efcd96-7044-41b1-a8b5-9207546046c2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.257408] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9fac0bd-f0d3-4ee4-9139-647b506fccc5 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1029.257704] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-01dce46e-bb3b-420b-9c81-a7af7c2a88bc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.267471] env[62965]: DEBUG oslo_vmware.api [None req-d9fac0bd-f0d3-4ee4-9139-647b506fccc5 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 1029.267471] env[62965]: value = "task-1390311" [ 1029.267471] env[62965]: _type = "Task" [ 1029.267471] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.278951] env[62965]: DEBUG oslo_vmware.api [None req-d9fac0bd-f0d3-4ee4-9139-647b506fccc5 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390311, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.306989] env[62965]: DEBUG nova.compute.manager [None req-667ab7c8-8108-40fe-ae71-ea26a63829ce tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1029.307451] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-667ab7c8-8108-40fe-ae71-ea26a63829ce tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1029.308253] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adde64f7-3bac-4d83-8550-768b11ed02ea {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.317889] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-667ab7c8-8108-40fe-ae71-ea26a63829ce tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1029.318440] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-05d7e287-e0e6-4d9e-bc3f-34de1e3ae074 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.325180] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8f1d0cb-c619-47c1-868b-895a9215229f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.329464] env[62965]: DEBUG oslo_vmware.api [None req-667ab7c8-8108-40fe-ae71-ea26a63829ce tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 1029.329464] env[62965]: value = "task-1390313" [ 1029.329464] env[62965]: _type = "Task" [ 1029.329464] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.336205] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceb1ea30-8912-4db5-bdd3-4eedf5ed9aff {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.342763] env[62965]: DEBUG oslo_vmware.api [None req-667ab7c8-8108-40fe-ae71-ea26a63829ce tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390313, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.371106] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5140827-ddc9-4d37-a0bb-9f4b1ee262d7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.379833] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53795970-35a7-4f3a-9e09-119ffc2b348e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.394951] env[62965]: DEBUG nova.compute.provider_tree [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1029.434153] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8332c061-e6d5-494e-ae5d-52f86cc0afc3 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.460137] env[62965]: DEBUG oslo_vmware.api [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390310, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.478903} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.462381] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] df03e327-959c-42af-85cd-a7af2ec4d0f7/df03e327-959c-42af-85cd-a7af2ec4d0f7.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1029.462611] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1029.463118] env[62965]: DEBUG nova.compute.manager [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Stashing vm_state: active {{(pid=62965) _prep_resize /opt/stack/nova/nova/compute/manager.py:5953}} [ 1029.465881] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-61915f12-193a-49ca-8a75-ef01a07776d6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.473542] env[62965]: DEBUG oslo_vmware.api [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 1029.473542] env[62965]: value = "task-1390314" [ 1029.473542] env[62965]: _type = "Task" [ 1029.473542] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.484965] env[62965]: DEBUG oslo_vmware.api [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390314, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.777225] env[62965]: DEBUG oslo_vmware.api [None req-d9fac0bd-f0d3-4ee4-9139-647b506fccc5 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390311, 'name': PowerOffVM_Task, 'duration_secs': 0.258501} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.777518] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9fac0bd-f0d3-4ee4-9139-647b506fccc5 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1029.777694] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-d9fac0bd-f0d3-4ee4-9139-647b506fccc5 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1029.777945] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-949fb57d-5330-4d25-9576-430f6fbb6405 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.838965] env[62965]: DEBUG oslo_vmware.api [None req-667ab7c8-8108-40fe-ae71-ea26a63829ce tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390313, 'name': PowerOffVM_Task, 'duration_secs': 0.243495} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.839255] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-667ab7c8-8108-40fe-ae71-ea26a63829ce tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1029.839436] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-667ab7c8-8108-40fe-ae71-ea26a63829ce tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1029.839673] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-330d8626-94ac-4dd6-b02e-01806e82b4ce {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.917028] env[62965]: ERROR nova.scheduler.client.report [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [req-b891583b-615a-4c55-b6c8-98f7cd322e3d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-b891583b-615a-4c55-b6c8-98f7cd322e3d"}]} [ 1029.933397] env[62965]: DEBUG nova.scheduler.client.report [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Refreshing inventories for resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 1029.948508] env[62965]: DEBUG nova.scheduler.client.report [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Updating ProviderTree inventory for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 1029.948748] env[62965]: DEBUG nova.compute.provider_tree [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1029.962227] env[62965]: DEBUG nova.scheduler.client.report [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Refreshing aggregate associations for resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8, aggregates: None {{(pid=62965) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 1029.978342] env[62965]: DEBUG nova.scheduler.client.report [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Refreshing trait associations for resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64 {{(pid=62965) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 1029.985601] env[62965]: DEBUG oslo_vmware.api [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390314, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070085} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.986560] env[62965]: DEBUG oslo_concurrency.lockutils [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.986752] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1029.987574] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b6a779d-6a27-403a-8fb2-a24cd76e52e8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.011538] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] df03e327-959c-42af-85cd-a7af2ec4d0f7/df03e327-959c-42af-85cd-a7af2ec4d0f7.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1030.014776] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f09ca80e-f52e-42b0-a9a9-970b8deae805 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.034522] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-d9fac0bd-f0d3-4ee4-9139-647b506fccc5 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1030.034669] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-d9fac0bd-f0d3-4ee4-9139-647b506fccc5 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Deleting contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1030.034854] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9fac0bd-f0d3-4ee4-9139-647b506fccc5 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Deleting the datastore file [datastore2] 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1030.035492] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-db8c640d-543f-4c6d-b861-f921b7ac9b26 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.040249] env[62965]: DEBUG oslo_vmware.api [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 1030.040249] env[62965]: value = "task-1390317" [ 1030.040249] env[62965]: _type = "Task" [ 1030.040249] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.044652] env[62965]: DEBUG oslo_vmware.api [None req-d9fac0bd-f0d3-4ee4-9139-647b506fccc5 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 1030.044652] env[62965]: value = "task-1390318" [ 1030.044652] env[62965]: _type = "Task" [ 1030.044652] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.053151] env[62965]: DEBUG oslo_vmware.api [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390317, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.057679] env[62965]: DEBUG oslo_vmware.api [None req-d9fac0bd-f0d3-4ee4-9139-647b506fccc5 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390318, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.063664] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-667ab7c8-8108-40fe-ae71-ea26a63829ce tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1030.063881] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-667ab7c8-8108-40fe-ae71-ea26a63829ce tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Deleting contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1030.064079] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-667ab7c8-8108-40fe-ae71-ea26a63829ce tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Deleting the datastore file [datastore2] 913c0536-5e39-4654-b083-2703a6aed328 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1030.064385] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bb755e0f-1814-4b38-89e1-161b5330e685 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.072747] env[62965]: DEBUG oslo_vmware.api [None req-667ab7c8-8108-40fe-ae71-ea26a63829ce tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for the task: (returnval){ [ 1030.072747] env[62965]: value = "task-1390319" [ 1030.072747] env[62965]: _type = "Task" [ 1030.072747] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.080305] env[62965]: DEBUG oslo_vmware.api [None req-667ab7c8-8108-40fe-ae71-ea26a63829ce tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390319, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.123930] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0745aa68-248b-46be-941d-e1e6cd7ec128 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.131320] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3807fbc5-6f35-4908-98c4-689100f16d62 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.161555] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8352b7e2-8f8b-470d-9091-796b28b87c49 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.168678] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcffe470-198d-49ef-b82c-c10aaf1e026c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.181413] env[62965]: DEBUG nova.compute.provider_tree [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1030.525905] env[62965]: DEBUG oslo_concurrency.lockutils [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "58c62ac1-203c-4370-8c94-420d655b7c08" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.526182] env[62965]: DEBUG oslo_concurrency.lockutils [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "58c62ac1-203c-4370-8c94-420d655b7c08" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.549706] env[62965]: DEBUG oslo_vmware.api [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390317, 'name': ReconfigVM_Task, 'duration_secs': 0.319785} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.552791] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Reconfigured VM instance instance-00000064 to attach disk [datastore2] df03e327-959c-42af-85cd-a7af2ec4d0f7/df03e327-959c-42af-85cd-a7af2ec4d0f7.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1030.553601] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a0c7e289-7033-483c-b6f2-08c9607db4cd {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.559264] env[62965]: DEBUG oslo_vmware.api [None req-d9fac0bd-f0d3-4ee4-9139-647b506fccc5 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390318, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144536} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.560293] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9fac0bd-f0d3-4ee4-9139-647b506fccc5 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1030.560483] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-d9fac0bd-f0d3-4ee4-9139-647b506fccc5 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Deleted contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1030.560662] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-d9fac0bd-f0d3-4ee4-9139-647b506fccc5 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1030.560833] env[62965]: INFO nova.compute.manager [None req-d9fac0bd-f0d3-4ee4-9139-647b506fccc5 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Took 1.31 seconds to destroy the instance on the hypervisor. [ 1030.561106] env[62965]: DEBUG oslo.service.loopingcall [None req-d9fac0bd-f0d3-4ee4-9139-647b506fccc5 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1030.561351] env[62965]: DEBUG oslo_vmware.api [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 1030.561351] env[62965]: value = "task-1390320" [ 1030.561351] env[62965]: _type = "Task" [ 1030.561351] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.561751] env[62965]: DEBUG nova.compute.manager [-] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1030.561848] env[62965]: DEBUG nova.network.neutron [-] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1030.570651] env[62965]: DEBUG oslo_vmware.api [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390320, 'name': Rename_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.580893] env[62965]: DEBUG oslo_vmware.api [None req-667ab7c8-8108-40fe-ae71-ea26a63829ce tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Task: {'id': task-1390319, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134729} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.581184] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-667ab7c8-8108-40fe-ae71-ea26a63829ce tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1030.581378] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-667ab7c8-8108-40fe-ae71-ea26a63829ce tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Deleted contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1030.581560] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-667ab7c8-8108-40fe-ae71-ea26a63829ce tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1030.581731] env[62965]: INFO nova.compute.manager [None req-667ab7c8-8108-40fe-ae71-ea26a63829ce tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Took 1.27 seconds to destroy the instance on the hypervisor. [ 1030.581966] env[62965]: DEBUG oslo.service.loopingcall [None req-667ab7c8-8108-40fe-ae71-ea26a63829ce tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1030.582181] env[62965]: DEBUG nova.compute.manager [-] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1030.582272] env[62965]: DEBUG nova.network.neutron [-] [instance: 913c0536-5e39-4654-b083-2703a6aed328] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1030.703741] env[62965]: ERROR nova.scheduler.client.report [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [req-da6938db-986a-445a-90f6-eff28b6b1994] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-da6938db-986a-445a-90f6-eff28b6b1994"}]} [ 1030.723610] env[62965]: DEBUG nova.scheduler.client.report [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Refreshing inventories for resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 1030.738171] env[62965]: DEBUG nova.scheduler.client.report [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Updating ProviderTree inventory for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 1030.738416] env[62965]: DEBUG nova.compute.provider_tree [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 174, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1030.752196] env[62965]: DEBUG nova.scheduler.client.report [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Refreshing aggregate associations for resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8, aggregates: None {{(pid=62965) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 1030.773036] env[62965]: DEBUG nova.scheduler.client.report [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Refreshing trait associations for resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64 {{(pid=62965) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 1030.859789] env[62965]: DEBUG nova.compute.manager [req-45fc85c1-339c-42ab-ae03-623839a41e0f req-b5369b0c-3c8a-4a0f-925f-713420c6db7c service nova] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Received event network-vif-deleted-6579674f-a74a-48ac-b7b8-accba556f17f {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1030.859789] env[62965]: INFO nova.compute.manager [req-45fc85c1-339c-42ab-ae03-623839a41e0f req-b5369b0c-3c8a-4a0f-925f-713420c6db7c service nova] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Neutron deleted interface 6579674f-a74a-48ac-b7b8-accba556f17f; detaching it from the instance and deleting it from the info cache [ 1030.859789] env[62965]: DEBUG nova.network.neutron [req-45fc85c1-339c-42ab-ae03-623839a41e0f req-b5369b0c-3c8a-4a0f-925f-713420c6db7c service nova] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.929044] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fa163da-c8af-4fe1-9e22-c936bddae532 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.937135] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74f6a833-e164-47b2-bb40-29c84b33e942 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.940956] env[62965]: DEBUG nova.compute.manager [req-aee3c59e-b507-4ddd-bca9-6c6b487971a0 req-50ab7406-67e2-4435-a310-d984a4f14c3e service nova] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Received event network-vif-deleted-9f81f2ea-9520-4f9c-a2f4-e6c3b1ab1216 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1030.941162] env[62965]: INFO nova.compute.manager [req-aee3c59e-b507-4ddd-bca9-6c6b487971a0 req-50ab7406-67e2-4435-a310-d984a4f14c3e service nova] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Neutron deleted interface 9f81f2ea-9520-4f9c-a2f4-e6c3b1ab1216; detaching it from the instance and deleting it from the info cache [ 1030.941343] env[62965]: DEBUG nova.network.neutron [req-aee3c59e-b507-4ddd-bca9-6c6b487971a0 req-50ab7406-67e2-4435-a310-d984a4f14c3e service nova] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.971065] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e844731e-646a-4f20-ba1e-45e1ceb8dba3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.979188] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67113789-6ab7-470e-a466-f9f57765a882 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.992855] env[62965]: DEBUG nova.compute.provider_tree [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1031.028070] env[62965]: DEBUG nova.compute.manager [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1031.073277] env[62965]: DEBUG oslo_vmware.api [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390320, 'name': Rename_Task, 'duration_secs': 0.135241} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.073554] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1031.073805] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3d3740c9-feea-446f-a9d6-c9f110b4b09a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.080407] env[62965]: DEBUG oslo_vmware.api [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 1031.080407] env[62965]: value = "task-1390321" [ 1031.080407] env[62965]: _type = "Task" [ 1031.080407] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.087866] env[62965]: DEBUG oslo_vmware.api [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390321, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.339784] env[62965]: DEBUG nova.network.neutron [-] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.365306] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-33fbe263-b159-46a3-965f-2a6f0434603a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.375012] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d00c9e0f-e689-4b0f-9d3c-76559da21f1b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.400536] env[62965]: DEBUG nova.compute.manager [req-45fc85c1-339c-42ab-ae03-623839a41e0f req-b5369b0c-3c8a-4a0f-925f-713420c6db7c service nova] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Detach interface failed, port_id=6579674f-a74a-48ac-b7b8-accba556f17f, reason: Instance 913c0536-5e39-4654-b083-2703a6aed328 could not be found. {{(pid=62965) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11294}} [ 1031.417226] env[62965]: DEBUG nova.network.neutron [-] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.443836] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1a6bbaaf-5103-45b1-8b6d-6078a6f14c62 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.453681] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83fda572-a6d8-4f2c-9881-6f85bdd8ffcf {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.481258] env[62965]: DEBUG nova.compute.manager [req-aee3c59e-b507-4ddd-bca9-6c6b487971a0 req-50ab7406-67e2-4435-a310-d984a4f14c3e service nova] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Detach interface failed, port_id=9f81f2ea-9520-4f9c-a2f4-e6c3b1ab1216, reason: Instance 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b could not be found. {{(pid=62965) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11294}} [ 1031.526370] env[62965]: DEBUG nova.scheduler.client.report [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Updated inventory for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with generation 128 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 1031.526721] env[62965]: DEBUG nova.compute.provider_tree [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Updating resource provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 generation from 128 to 129 during operation: update_inventory {{(pid=62965) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1031.526826] env[62965]: DEBUG nova.compute.provider_tree [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Updating inventory in ProviderTree for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1031.549121] env[62965]: DEBUG oslo_concurrency.lockutils [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.590470] env[62965]: DEBUG oslo_vmware.api [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390321, 'name': PowerOnVM_Task, 'duration_secs': 0.427267} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.590693] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1031.590894] env[62965]: INFO nova.compute.manager [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Took 7.09 seconds to spawn the instance on the hypervisor. [ 1031.591128] env[62965]: DEBUG nova.compute.manager [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1031.591864] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0be6ca7d-9790-47e7-86c3-edf7cb5da957 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.842502] env[62965]: INFO nova.compute.manager [-] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Took 1.26 seconds to deallocate network for instance. [ 1031.919758] env[62965]: INFO nova.compute.manager [-] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Took 1.36 seconds to deallocate network for instance. [ 1032.032112] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.352s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.034999] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8332c061-e6d5-494e-ae5d-52f86cc0afc3 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.600s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.035483] env[62965]: DEBUG nova.objects.instance [None req-8332c061-e6d5-494e-ae5d-52f86cc0afc3 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lazy-loading 'resources' on Instance uuid 8ae1fde5-e94e-43ed-816a-9e22633dfdd1 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1032.052432] env[62965]: INFO nova.scheduler.client.report [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Deleted allocations for instance b59d6023-a047-4b32-bd18-1c7cc05f67d7 [ 1032.108152] env[62965]: INFO nova.compute.manager [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Took 18.04 seconds to build instance. [ 1032.348714] env[62965]: DEBUG oslo_concurrency.lockutils [None req-667ab7c8-8108-40fe-ae71-ea26a63829ce tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.425661] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d9fac0bd-f0d3-4ee4-9139-647b506fccc5 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.562611] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ea5580dc-1d2a-494d-a1f1-40c765d8011b tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "b59d6023-a047-4b32-bd18-1c7cc05f67d7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.272s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.610900] env[62965]: DEBUG oslo_concurrency.lockutils [None req-6d61de8d-e0f5-4a42-911c-248fb05b3192 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "df03e327-959c-42af-85cd-a7af2ec4d0f7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.555s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.688967] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef3b50e6-e84b-4a99-998e-3a4f128e7b30 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.698544] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32af616f-fdb5-412d-9af6-e97ca437e31e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.731682] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2fcf1ee-8ffd-494e-ae8b-ea37829c24b2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.740022] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-087a7022-934f-4695-9df3-1b065cfa36eb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.753804] env[62965]: DEBUG nova.compute.provider_tree [None req-8332c061-e6d5-494e-ae5d-52f86cc0afc3 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1033.256857] env[62965]: DEBUG nova.scheduler.client.report [None req-8332c061-e6d5-494e-ae5d-52f86cc0afc3 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1033.766023] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8332c061-e6d5-494e-ae5d-52f86cc0afc3 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.729s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.766023] env[62965]: DEBUG oslo_concurrency.lockutils [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 3.779s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.792299] env[62965]: INFO nova.scheduler.client.report [None req-8332c061-e6d5-494e-ae5d-52f86cc0afc3 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Deleted allocations for instance 8ae1fde5-e94e-43ed-816a-9e22633dfdd1 [ 1034.246032] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4fa857a1-4a4c-4f0d-8ac0-6abcd392cc99 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "df03e327-959c-42af-85cd-a7af2ec4d0f7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.246032] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4fa857a1-4a4c-4f0d-8ac0-6abcd392cc99 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "df03e327-959c-42af-85cd-a7af2ec4d0f7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.246032] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4fa857a1-4a4c-4f0d-8ac0-6abcd392cc99 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "df03e327-959c-42af-85cd-a7af2ec4d0f7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.246032] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4fa857a1-4a4c-4f0d-8ac0-6abcd392cc99 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "df03e327-959c-42af-85cd-a7af2ec4d0f7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.246032] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4fa857a1-4a4c-4f0d-8ac0-6abcd392cc99 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "df03e327-959c-42af-85cd-a7af2ec4d0f7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.247610] env[62965]: INFO nova.compute.manager [None req-4fa857a1-4a4c-4f0d-8ac0-6abcd392cc99 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Terminating instance [ 1034.272467] env[62965]: INFO nova.compute.claims [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1034.303833] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8332c061-e6d5-494e-ae5d-52f86cc0afc3 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "8ae1fde5-e94e-43ed-816a-9e22633dfdd1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.722s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.370322] env[62965]: DEBUG oslo_concurrency.lockutils [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquiring lock "73a7ca0e-b1d3-4464-969b-5111d2506a8b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.370439] env[62965]: DEBUG oslo_concurrency.lockutils [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "73a7ca0e-b1d3-4464-969b-5111d2506a8b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.754400] env[62965]: DEBUG nova.compute.manager [None req-4fa857a1-4a4c-4f0d-8ac0-6abcd392cc99 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1034.754590] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-4fa857a1-4a4c-4f0d-8ac0-6abcd392cc99 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1034.755539] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9e94161-9552-42cb-b47b-51a780ad658d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.763394] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fa857a1-4a4c-4f0d-8ac0-6abcd392cc99 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1034.763631] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-544c20d8-e0bd-4164-9a40-6a5cfbad31b1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.770110] env[62965]: DEBUG oslo_vmware.api [None req-4fa857a1-4a4c-4f0d-8ac0-6abcd392cc99 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 1034.770110] env[62965]: value = "task-1390322" [ 1034.770110] env[62965]: _type = "Task" [ 1034.770110] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.778149] env[62965]: DEBUG oslo_vmware.api [None req-4fa857a1-4a4c-4f0d-8ac0-6abcd392cc99 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390322, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.785288] env[62965]: INFO nova.compute.resource_tracker [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Updating resource usage from migration a1de2151-bfa9-4136-9a84-6811eb0921cc [ 1034.873114] env[62965]: DEBUG nova.compute.manager [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1034.912174] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7639a841-e677-426e-a256-d65e8c0470d3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.921058] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6baf46ee-ec8f-4b05-8e6d-41f96430eae2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.954745] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4823d79-118d-4bc0-899f-13b6740d57de {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.962554] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dbf4c64-fe57-48ae-955f-68ff0dd5888d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.976289] env[62965]: DEBUG nova.compute.provider_tree [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1035.062244] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "88e45c29-4d21-4f91-92f4-fdebaf905ccb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.062506] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "88e45c29-4d21-4f91-92f4-fdebaf905ccb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.279278] env[62965]: DEBUG oslo_vmware.api [None req-4fa857a1-4a4c-4f0d-8ac0-6abcd392cc99 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390322, 'name': PowerOffVM_Task, 'duration_secs': 0.178816} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.279567] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fa857a1-4a4c-4f0d-8ac0-6abcd392cc99 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1035.279748] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-4fa857a1-4a4c-4f0d-8ac0-6abcd392cc99 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1035.280007] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b79e41f7-7493-4751-b119-5402a3727c5f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.342250] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-4fa857a1-4a4c-4f0d-8ac0-6abcd392cc99 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1035.342483] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-4fa857a1-4a4c-4f0d-8ac0-6abcd392cc99 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Deleting contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1035.342667] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-4fa857a1-4a4c-4f0d-8ac0-6abcd392cc99 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Deleting the datastore file [datastore2] df03e327-959c-42af-85cd-a7af2ec4d0f7 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1035.342943] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fe88515b-4e1c-4d81-8ce0-6078e53be67a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.350454] env[62965]: DEBUG oslo_vmware.api [None req-4fa857a1-4a4c-4f0d-8ac0-6abcd392cc99 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for the task: (returnval){ [ 1035.350454] env[62965]: value = "task-1390324" [ 1035.350454] env[62965]: _type = "Task" [ 1035.350454] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.357941] env[62965]: DEBUG oslo_vmware.api [None req-4fa857a1-4a4c-4f0d-8ac0-6abcd392cc99 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390324, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.392896] env[62965]: DEBUG oslo_concurrency.lockutils [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.479530] env[62965]: DEBUG nova.scheduler.client.report [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1035.565139] env[62965]: DEBUG nova.compute.manager [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1035.864674] env[62965]: DEBUG oslo_vmware.api [None req-4fa857a1-4a4c-4f0d-8ac0-6abcd392cc99 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Task: {'id': task-1390324, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13051} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.864674] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-4fa857a1-4a4c-4f0d-8ac0-6abcd392cc99 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1035.864674] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-4fa857a1-4a4c-4f0d-8ac0-6abcd392cc99 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Deleted contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1035.864674] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-4fa857a1-4a4c-4f0d-8ac0-6abcd392cc99 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1035.864674] env[62965]: INFO nova.compute.manager [None req-4fa857a1-4a4c-4f0d-8ac0-6abcd392cc99 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1035.864674] env[62965]: DEBUG oslo.service.loopingcall [None req-4fa857a1-4a4c-4f0d-8ac0-6abcd392cc99 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1035.864674] env[62965]: DEBUG nova.compute.manager [-] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1035.864674] env[62965]: DEBUG nova.network.neutron [-] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1035.985934] env[62965]: DEBUG oslo_concurrency.lockutils [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.219s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.985934] env[62965]: INFO nova.compute.manager [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Migrating [ 1035.991234] env[62965]: DEBUG oslo_concurrency.lockutils [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.442s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.993031] env[62965]: INFO nova.compute.claims [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1036.092797] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.118353] env[62965]: DEBUG nova.compute.manager [req-2ae586aa-211f-494a-a30e-c2bf57119739 req-ffa66762-a688-4f9c-ab8a-1bd89a49d5d2 service nova] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Received event network-vif-deleted-0ecbf02c-8b40-41fc-a928-d81ec8eeb2a0 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1036.118353] env[62965]: INFO nova.compute.manager [req-2ae586aa-211f-494a-a30e-c2bf57119739 req-ffa66762-a688-4f9c-ab8a-1bd89a49d5d2 service nova] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Neutron deleted interface 0ecbf02c-8b40-41fc-a928-d81ec8eeb2a0; detaching it from the instance and deleting it from the info cache [ 1036.118353] env[62965]: DEBUG nova.network.neutron [req-2ae586aa-211f-494a-a30e-c2bf57119739 req-ffa66762-a688-4f9c-ab8a-1bd89a49d5d2 service nova] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.502502] env[62965]: DEBUG oslo_concurrency.lockutils [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "refresh_cache-c2448b76-1553-4b68-a731-0bb0967f4c1d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1036.502705] env[62965]: DEBUG oslo_concurrency.lockutils [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquired lock "refresh_cache-c2448b76-1553-4b68-a731-0bb0967f4c1d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1036.502872] env[62965]: DEBUG nova.network.neutron [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1036.590484] env[62965]: DEBUG nova.network.neutron [-] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.620461] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a8aee928-592e-4c31-b213-16ad43ec9a87 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.633645] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0e236ce-8bbc-4c0e-b124-4f00c8e9e0a6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.664921] env[62965]: DEBUG nova.compute.manager [req-2ae586aa-211f-494a-a30e-c2bf57119739 req-ffa66762-a688-4f9c-ab8a-1bd89a49d5d2 service nova] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Detach interface failed, port_id=0ecbf02c-8b40-41fc-a928-d81ec8eeb2a0, reason: Instance df03e327-959c-42af-85cd-a7af2ec4d0f7 could not be found. {{(pid=62965) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11294}} [ 1037.093140] env[62965]: INFO nova.compute.manager [-] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Took 1.23 seconds to deallocate network for instance. [ 1037.132286] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76fdee79-7b8b-4b50-a09d-06b62c215710 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.140889] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91686d09-e000-4442-b05f-34e917c3237b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.174992] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bb2648d-61b7-4aa9-aa04-138a89e371d5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.184376] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2e5338a-b52b-43a6-afcc-c906911b0c7d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.199675] env[62965]: DEBUG nova.compute.provider_tree [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1037.248711] env[62965]: DEBUG nova.network.neutron [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Updating instance_info_cache with network_info: [{"id": "57672e8b-1df1-463d-b9b3-8290667d9fd2", "address": "fa:16:3e:53:46:47", "network": {"id": "4649fed7-c2ca-43a9-a183-7c99423da726", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1978698173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fab418e2fe1420793517663574b43bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43ad01d2-c7dd-453c-a929-8ad76294d13c", "external-id": "nsx-vlan-transportzone-176", "segmentation_id": 176, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57672e8b-1d", "ovs_interfaceid": "57672e8b-1df1-463d-b9b3-8290667d9fd2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.599195] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4fa857a1-4a4c-4f0d-8ac0-6abcd392cc99 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.703588] env[62965]: DEBUG nova.scheduler.client.report [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1037.750667] env[62965]: DEBUG oslo_concurrency.lockutils [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Releasing lock "refresh_cache-c2448b76-1553-4b68-a731-0bb0967f4c1d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.208675] env[62965]: DEBUG oslo_concurrency.lockutils [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.217s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.209232] env[62965]: DEBUG nova.compute.manager [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1038.211771] env[62965]: DEBUG oslo_concurrency.lockutils [None req-667ab7c8-8108-40fe-ae71-ea26a63829ce tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.863s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.211993] env[62965]: DEBUG nova.objects.instance [None req-667ab7c8-8108-40fe-ae71-ea26a63829ce tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lazy-loading 'resources' on Instance uuid 913c0536-5e39-4654-b083-2703a6aed328 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1038.715275] env[62965]: DEBUG nova.compute.utils [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1038.719437] env[62965]: DEBUG nova.compute.manager [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1038.719611] env[62965]: DEBUG nova.network.neutron [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1038.762783] env[62965]: DEBUG nova.policy [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '97ee4fdb7cdb45829c7fe694e883da8c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1ed71548c085499981fa1b102d660368', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 1038.842654] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c454cc6-ef03-4686-9a76-cb3f0648905b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.850313] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-442973a0-0a61-469c-ab56-9c9469d3a209 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.882142] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6f24988-6728-4cda-b9a9-f6d03e670080 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.889541] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23060240-dc61-446d-a89e-6b243f131964 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.902978] env[62965]: DEBUG nova.compute.provider_tree [None req-667ab7c8-8108-40fe-ae71-ea26a63829ce tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1039.016588] env[62965]: DEBUG nova.network.neutron [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Successfully created port: d67556e3-a304-4a17-9d28-b275448ee50c {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1039.220058] env[62965]: DEBUG nova.compute.manager [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1039.269989] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e8681df-4056-4804-8dc0-9382b0b9e209 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.288761] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Updating instance 'c2448b76-1553-4b68-a731-0bb0967f4c1d' progress to 0 {{(pid=62965) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1039.407052] env[62965]: DEBUG nova.scheduler.client.report [None req-667ab7c8-8108-40fe-ae71-ea26a63829ce tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1039.794835] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1039.795127] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-346b4add-36ff-4b11-9cf3-d0b8bae8fed3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.802279] env[62965]: DEBUG oslo_vmware.api [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1039.802279] env[62965]: value = "task-1390325" [ 1039.802279] env[62965]: _type = "Task" [ 1039.802279] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.809747] env[62965]: DEBUG oslo_vmware.api [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390325, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.911298] env[62965]: DEBUG oslo_concurrency.lockutils [None req-667ab7c8-8108-40fe-ae71-ea26a63829ce tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.699s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.913633] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d9fac0bd-f0d3-4ee4-9139-647b506fccc5 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.488s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.913885] env[62965]: DEBUG nova.objects.instance [None req-d9fac0bd-f0d3-4ee4-9139-647b506fccc5 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lazy-loading 'resources' on Instance uuid 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1039.934783] env[62965]: INFO nova.scheduler.client.report [None req-667ab7c8-8108-40fe-ae71-ea26a63829ce tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Deleted allocations for instance 913c0536-5e39-4654-b083-2703a6aed328 [ 1040.231282] env[62965]: DEBUG nova.compute.manager [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1040.255423] env[62965]: DEBUG nova.virt.hardware [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1040.255690] env[62965]: DEBUG nova.virt.hardware [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1040.255853] env[62965]: DEBUG nova.virt.hardware [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1040.256054] env[62965]: DEBUG nova.virt.hardware [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1040.256212] env[62965]: DEBUG nova.virt.hardware [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1040.256366] env[62965]: DEBUG nova.virt.hardware [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1040.256578] env[62965]: DEBUG nova.virt.hardware [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1040.256733] env[62965]: DEBUG nova.virt.hardware [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1040.256903] env[62965]: DEBUG nova.virt.hardware [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1040.257083] env[62965]: DEBUG nova.virt.hardware [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1040.257267] env[62965]: DEBUG nova.virt.hardware [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1040.258137] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99cefdda-be31-4433-a661-7a3788339765 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.265706] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7878f0a1-d73a-4162-8c15-d8c6cbd12f6e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.311479] env[62965]: DEBUG oslo_vmware.api [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390325, 'name': PowerOffVM_Task, 'duration_secs': 0.179963} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.311688] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1040.311879] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Updating instance 'c2448b76-1553-4b68-a731-0bb0967f4c1d' progress to 17 {{(pid=62965) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1040.378441] env[62965]: DEBUG nova.compute.manager [req-97385bfe-174a-485b-8b82-8e673e0ff050 req-4a167c9e-4ace-47c3-a259-974abae0a583 service nova] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Received event network-vif-plugged-d67556e3-a304-4a17-9d28-b275448ee50c {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1040.378669] env[62965]: DEBUG oslo_concurrency.lockutils [req-97385bfe-174a-485b-8b82-8e673e0ff050 req-4a167c9e-4ace-47c3-a259-974abae0a583 service nova] Acquiring lock "58c62ac1-203c-4370-8c94-420d655b7c08-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.378884] env[62965]: DEBUG oslo_concurrency.lockutils [req-97385bfe-174a-485b-8b82-8e673e0ff050 req-4a167c9e-4ace-47c3-a259-974abae0a583 service nova] Lock "58c62ac1-203c-4370-8c94-420d655b7c08-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.379072] env[62965]: DEBUG oslo_concurrency.lockutils [req-97385bfe-174a-485b-8b82-8e673e0ff050 req-4a167c9e-4ace-47c3-a259-974abae0a583 service nova] Lock "58c62ac1-203c-4370-8c94-420d655b7c08-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.379248] env[62965]: DEBUG nova.compute.manager [req-97385bfe-174a-485b-8b82-8e673e0ff050 req-4a167c9e-4ace-47c3-a259-974abae0a583 service nova] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] No waiting events found dispatching network-vif-plugged-d67556e3-a304-4a17-9d28-b275448ee50c {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1040.379425] env[62965]: WARNING nova.compute.manager [req-97385bfe-174a-485b-8b82-8e673e0ff050 req-4a167c9e-4ace-47c3-a259-974abae0a583 service nova] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Received unexpected event network-vif-plugged-d67556e3-a304-4a17-9d28-b275448ee50c for instance with vm_state building and task_state spawning. [ 1040.446569] env[62965]: DEBUG oslo_concurrency.lockutils [None req-667ab7c8-8108-40fe-ae71-ea26a63829ce tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "913c0536-5e39-4654-b083-2703a6aed328" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.645s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.541998] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66615c0e-b904-4736-9f96-871e7772ca88 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.549510] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85c78990-7b48-46dc-9251-5410e3eecbb2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.578910] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bc2f64f-edb1-497e-9149-d3200c5e75b0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.586023] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3af7879f-ae25-42ef-a43c-5f2bd6c7c8e8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.600148] env[62965]: DEBUG nova.compute.provider_tree [None req-d9fac0bd-f0d3-4ee4-9139-647b506fccc5 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1040.818913] env[62965]: DEBUG nova.virt.hardware [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:24Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1040.819196] env[62965]: DEBUG nova.virt.hardware [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1040.819374] env[62965]: DEBUG nova.virt.hardware [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1040.819562] env[62965]: DEBUG nova.virt.hardware [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1040.819712] env[62965]: DEBUG nova.virt.hardware [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1040.819863] env[62965]: DEBUG nova.virt.hardware [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1040.820086] env[62965]: DEBUG nova.virt.hardware [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1040.820257] env[62965]: DEBUG nova.virt.hardware [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1040.820429] env[62965]: DEBUG nova.virt.hardware [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1040.820594] env[62965]: DEBUG nova.virt.hardware [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1040.820769] env[62965]: DEBUG nova.virt.hardware [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1040.825648] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4d7f3811-a5e1-4f5f-87aa-2a907b7a822f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.842504] env[62965]: DEBUG oslo_vmware.api [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1040.842504] env[62965]: value = "task-1390326" [ 1040.842504] env[62965]: _type = "Task" [ 1040.842504] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.850151] env[62965]: DEBUG oslo_vmware.api [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390326, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.930753] env[62965]: DEBUG nova.network.neutron [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Successfully updated port: d67556e3-a304-4a17-9d28-b275448ee50c {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1040.956780] env[62965]: DEBUG nova.compute.manager [req-d976783a-732c-40c6-b4d1-95a55742dc05 req-d779d40a-1505-4a0e-9b37-bd4888f05efd service nova] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Received event network-changed-d67556e3-a304-4a17-9d28-b275448ee50c {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1040.956957] env[62965]: DEBUG nova.compute.manager [req-d976783a-732c-40c6-b4d1-95a55742dc05 req-d779d40a-1505-4a0e-9b37-bd4888f05efd service nova] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Refreshing instance network info cache due to event network-changed-d67556e3-a304-4a17-9d28-b275448ee50c. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 1040.957216] env[62965]: DEBUG oslo_concurrency.lockutils [req-d976783a-732c-40c6-b4d1-95a55742dc05 req-d779d40a-1505-4a0e-9b37-bd4888f05efd service nova] Acquiring lock "refresh_cache-58c62ac1-203c-4370-8c94-420d655b7c08" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.957373] env[62965]: DEBUG oslo_concurrency.lockutils [req-d976783a-732c-40c6-b4d1-95a55742dc05 req-d779d40a-1505-4a0e-9b37-bd4888f05efd service nova] Acquired lock "refresh_cache-58c62ac1-203c-4370-8c94-420d655b7c08" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.957539] env[62965]: DEBUG nova.network.neutron [req-d976783a-732c-40c6-b4d1-95a55742dc05 req-d779d40a-1505-4a0e-9b37-bd4888f05efd service nova] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Refreshing network info cache for port d67556e3-a304-4a17-9d28-b275448ee50c {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1041.106066] env[62965]: DEBUG nova.scheduler.client.report [None req-d9fac0bd-f0d3-4ee4-9139-647b506fccc5 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1041.352374] env[62965]: DEBUG oslo_vmware.api [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390326, 'name': ReconfigVM_Task, 'duration_secs': 0.182814} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.352707] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Updating instance 'c2448b76-1553-4b68-a731-0bb0967f4c1d' progress to 33 {{(pid=62965) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1041.436531] env[62965]: DEBUG oslo_concurrency.lockutils [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "refresh_cache-58c62ac1-203c-4370-8c94-420d655b7c08" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1041.488865] env[62965]: DEBUG nova.network.neutron [req-d976783a-732c-40c6-b4d1-95a55742dc05 req-d779d40a-1505-4a0e-9b37-bd4888f05efd service nova] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1041.576181] env[62965]: DEBUG nova.network.neutron [req-d976783a-732c-40c6-b4d1-95a55742dc05 req-d779d40a-1505-4a0e-9b37-bd4888f05efd service nova] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.608741] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d9fac0bd-f0d3-4ee4-9139-647b506fccc5 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.695s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.610895] env[62965]: DEBUG oslo_concurrency.lockutils [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.218s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.612548] env[62965]: INFO nova.compute.claims [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1041.627964] env[62965]: INFO nova.scheduler.client.report [None req-d9fac0bd-f0d3-4ee4-9139-647b506fccc5 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Deleted allocations for instance 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b [ 1041.859071] env[62965]: DEBUG nova.virt.hardware [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1041.859381] env[62965]: DEBUG nova.virt.hardware [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1041.859565] env[62965]: DEBUG nova.virt.hardware [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1041.859758] env[62965]: DEBUG nova.virt.hardware [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1041.859911] env[62965]: DEBUG nova.virt.hardware [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1041.860076] env[62965]: DEBUG nova.virt.hardware [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1041.860289] env[62965]: DEBUG nova.virt.hardware [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1041.860455] env[62965]: DEBUG nova.virt.hardware [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1041.860623] env[62965]: DEBUG nova.virt.hardware [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1041.860785] env[62965]: DEBUG nova.virt.hardware [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1041.860959] env[62965]: DEBUG nova.virt.hardware [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1041.866218] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Reconfiguring VM instance instance-00000042 to detach disk 2000 {{(pid=62965) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1041.866521] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8e14cdbc-2dd0-4996-b71d-caa024fbcfdb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.885810] env[62965]: DEBUG oslo_vmware.api [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1041.885810] env[62965]: value = "task-1390327" [ 1041.885810] env[62965]: _type = "Task" [ 1041.885810] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.893558] env[62965]: DEBUG oslo_vmware.api [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390327, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.079359] env[62965]: DEBUG oslo_concurrency.lockutils [req-d976783a-732c-40c6-b4d1-95a55742dc05 req-d779d40a-1505-4a0e-9b37-bd4888f05efd service nova] Releasing lock "refresh_cache-58c62ac1-203c-4370-8c94-420d655b7c08" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.079791] env[62965]: DEBUG oslo_concurrency.lockutils [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquired lock "refresh_cache-58c62ac1-203c-4370-8c94-420d655b7c08" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.079949] env[62965]: DEBUG nova.network.neutron [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1042.135306] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d9fac0bd-f0d3-4ee4-9139-647b506fccc5 tempest-MultipleCreateTestJSON-155465349 tempest-MultipleCreateTestJSON-155465349-project-member] Lock "9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.398s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.395801] env[62965]: DEBUG oslo_vmware.api [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390327, 'name': ReconfigVM_Task, 'duration_secs': 0.154487} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.396120] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Reconfigured VM instance instance-00000042 to detach disk 2000 {{(pid=62965) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1042.396871] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d2c1433-f0a3-4b70-a227-11028e3bf8f9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.418461] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] c2448b76-1553-4b68-a731-0bb0967f4c1d/c2448b76-1553-4b68-a731-0bb0967f4c1d.vmdk or device None with type thin {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1042.418612] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3db291ad-a1ba-47e7-8ab0-a8dfef29a08f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.437028] env[62965]: DEBUG oslo_vmware.api [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1042.437028] env[62965]: value = "task-1390328" [ 1042.437028] env[62965]: _type = "Task" [ 1042.437028] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.445120] env[62965]: DEBUG oslo_vmware.api [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390328, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.623070] env[62965]: DEBUG nova.network.neutron [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1042.729764] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1a40d25-2d9f-4bf9-9ddb-7c842b053087 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.739274] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dda6cf6-1eb7-4fb9-b756-dd7902b7feb8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.768641] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1257d638-141e-4daa-981f-c0b24f33bed4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.771863] env[62965]: DEBUG nova.network.neutron [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Updating instance_info_cache with network_info: [{"id": "d67556e3-a304-4a17-9d28-b275448ee50c", "address": "fa:16:3e:fb:75:8b", "network": {"id": "a0a5e605-d580-4b7b-b3c0-5c7395bd5a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1679618017-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ed71548c085499981fa1b102d660368", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd67556e3-a3", "ovs_interfaceid": "d67556e3-a304-4a17-9d28-b275448ee50c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.777803] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae916458-28be-4808-bf76-2c4b681b0768 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.791969] env[62965]: DEBUG nova.compute.provider_tree [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1042.948503] env[62965]: DEBUG oslo_vmware.api [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390328, 'name': ReconfigVM_Task, 'duration_secs': 0.24556} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.948831] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Reconfigured VM instance instance-00000042 to attach disk [datastore1] c2448b76-1553-4b68-a731-0bb0967f4c1d/c2448b76-1553-4b68-a731-0bb0967f4c1d.vmdk or device None with type thin {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1042.949189] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Updating instance 'c2448b76-1553-4b68-a731-0bb0967f4c1d' progress to 50 {{(pid=62965) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1043.274862] env[62965]: DEBUG oslo_concurrency.lockutils [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Releasing lock "refresh_cache-58c62ac1-203c-4370-8c94-420d655b7c08" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.275218] env[62965]: DEBUG nova.compute.manager [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Instance network_info: |[{"id": "d67556e3-a304-4a17-9d28-b275448ee50c", "address": "fa:16:3e:fb:75:8b", "network": {"id": "a0a5e605-d580-4b7b-b3c0-5c7395bd5a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1679618017-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ed71548c085499981fa1b102d660368", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd67556e3-a3", "ovs_interfaceid": "d67556e3-a304-4a17-9d28-b275448ee50c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1043.275737] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fb:75:8b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0cd5d325-3053-407e-a4ee-f627e82a23f9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd67556e3-a304-4a17-9d28-b275448ee50c', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1043.284860] env[62965]: DEBUG oslo.service.loopingcall [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1043.285122] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1043.285421] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d61256e3-5c75-485b-8dc2-131c30558a77 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.301399] env[62965]: DEBUG nova.scheduler.client.report [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1043.310976] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1043.310976] env[62965]: value = "task-1390329" [ 1043.310976] env[62965]: _type = "Task" [ 1043.310976] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.319374] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390329, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.456860] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ba85572-fdc6-493c-95ed-418285c0b64a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.480744] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45807544-915b-4b30-b41d-160ad646908f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.499515] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Updating instance 'c2448b76-1553-4b68-a731-0bb0967f4c1d' progress to 67 {{(pid=62965) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1043.807970] env[62965]: DEBUG oslo_concurrency.lockutils [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.197s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.808525] env[62965]: DEBUG nova.compute.manager [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1043.811732] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.719s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.813660] env[62965]: INFO nova.compute.claims [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1043.825802] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390329, 'name': CreateVM_Task, 'duration_secs': 0.450712} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.825960] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1043.828036] env[62965]: DEBUG oslo_concurrency.lockutils [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.829772] env[62965]: DEBUG oslo_concurrency.lockutils [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.829772] env[62965]: DEBUG oslo_concurrency.lockutils [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1043.829772] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5194eed7-b01d-43f9-ac8a-4eb829f49dca {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.833986] env[62965]: DEBUG oslo_vmware.api [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 1043.833986] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]523e7277-3782-7a81-dd90-fb65250791fd" [ 1043.833986] env[62965]: _type = "Task" [ 1043.833986] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.842776] env[62965]: DEBUG oslo_vmware.api [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]523e7277-3782-7a81-dd90-fb65250791fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.063341] env[62965]: DEBUG nova.network.neutron [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Port 57672e8b-1df1-463d-b9b3-8290667d9fd2 binding to destination host cpu-1 is already ACTIVE {{(pid=62965) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1044.314709] env[62965]: DEBUG nova.compute.utils [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1044.316117] env[62965]: DEBUG nova.compute.manager [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1044.316288] env[62965]: DEBUG nova.network.neutron [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1044.344462] env[62965]: DEBUG oslo_vmware.api [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]523e7277-3782-7a81-dd90-fb65250791fd, 'name': SearchDatastore_Task, 'duration_secs': 0.010058} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.344701] env[62965]: DEBUG oslo_concurrency.lockutils [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.344934] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1044.345192] env[62965]: DEBUG oslo_concurrency.lockutils [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1044.345345] env[62965]: DEBUG oslo_concurrency.lockutils [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.345523] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1044.345782] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-737b39c4-1d92-46b8-9768-6a50b824d801 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.354248] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1044.354433] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1044.355140] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a6aba449-d0c5-48b3-af48-9c0056877255 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.359813] env[62965]: DEBUG oslo_vmware.api [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 1044.359813] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52adde7e-5d39-bcda-6400-1132692597ba" [ 1044.359813] env[62965]: _type = "Task" [ 1044.359813] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.367257] env[62965]: DEBUG oslo_vmware.api [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52adde7e-5d39-bcda-6400-1132692597ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.372833] env[62965]: DEBUG nova.policy [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd2db1d35b31d477ca92c4ba7a3c53cd3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd91c99b14018469e8998fd1ccc0fab4f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 1044.627169] env[62965]: DEBUG nova.network.neutron [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Successfully created port: 5f939a36-bc5f-4fd7-b29e-011f72b3256c {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1044.821291] env[62965]: DEBUG nova.compute.manager [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1044.871704] env[62965]: DEBUG oslo_vmware.api [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52adde7e-5d39-bcda-6400-1132692597ba, 'name': SearchDatastore_Task, 'duration_secs': 0.007821} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.872572] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b729c30-43dc-4754-9cab-cba39817a1e5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.877472] env[62965]: DEBUG oslo_vmware.api [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 1044.877472] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52802d73-4e46-a4dd-d696-ab5fe5bcd646" [ 1044.877472] env[62965]: _type = "Task" [ 1044.877472] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.888458] env[62965]: DEBUG oslo_vmware.api [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52802d73-4e46-a4dd-d696-ab5fe5bcd646, 'name': SearchDatastore_Task, 'duration_secs': 0.008556} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.888836] env[62965]: DEBUG oslo_concurrency.lockutils [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.888957] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 58c62ac1-203c-4370-8c94-420d655b7c08/58c62ac1-203c-4370-8c94-420d655b7c08.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1044.889366] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f113d410-1011-4776-8f45-93e33d905e31 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.895103] env[62965]: DEBUG oslo_vmware.api [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 1044.895103] env[62965]: value = "task-1390330" [ 1044.895103] env[62965]: _type = "Task" [ 1044.895103] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.902531] env[62965]: DEBUG oslo_vmware.api [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390330, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.933412] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbba2dbb-ff7f-4778-a72c-260e44c3e438 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.940034] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bda0888-5551-489c-afa9-b1459abe4d81 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.975731] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e75001b5-c29d-419e-b05d-6138d5091757 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.983927] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d52920e6-df07-4838-8005-5b058b2dd63b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.999183] env[62965]: DEBUG nova.compute.provider_tree [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1045.087711] env[62965]: DEBUG oslo_concurrency.lockutils [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "c2448b76-1553-4b68-a731-0bb0967f4c1d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.087711] env[62965]: DEBUG oslo_concurrency.lockutils [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "c2448b76-1553-4b68-a731-0bb0967f4c1d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.087914] env[62965]: DEBUG oslo_concurrency.lockutils [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "c2448b76-1553-4b68-a731-0bb0967f4c1d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.405041] env[62965]: DEBUG oslo_vmware.api [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390330, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.45325} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.405041] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 58c62ac1-203c-4370-8c94-420d655b7c08/58c62ac1-203c-4370-8c94-420d655b7c08.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1045.405209] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1045.405540] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-48e381e6-bb79-4684-8834-fd237cf16e48 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.412485] env[62965]: DEBUG oslo_vmware.api [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 1045.412485] env[62965]: value = "task-1390331" [ 1045.412485] env[62965]: _type = "Task" [ 1045.412485] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.420365] env[62965]: DEBUG oslo_vmware.api [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390331, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.502687] env[62965]: DEBUG nova.scheduler.client.report [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1045.833839] env[62965]: DEBUG nova.compute.manager [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1045.857495] env[62965]: DEBUG nova.virt.hardware [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1045.857764] env[62965]: DEBUG nova.virt.hardware [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1045.857927] env[62965]: DEBUG nova.virt.hardware [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1045.858131] env[62965]: DEBUG nova.virt.hardware [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1045.858287] env[62965]: DEBUG nova.virt.hardware [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1045.858442] env[62965]: DEBUG nova.virt.hardware [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1045.858649] env[62965]: DEBUG nova.virt.hardware [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1045.858811] env[62965]: DEBUG nova.virt.hardware [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1045.858983] env[62965]: DEBUG nova.virt.hardware [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1045.859168] env[62965]: DEBUG nova.virt.hardware [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1045.859385] env[62965]: DEBUG nova.virt.hardware [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1045.860533] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-943e8d29-6f43-4bc7-b246-f501e787dc70 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.868059] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7a17c01-3813-4c07-a50b-05947c572254 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.920404] env[62965]: DEBUG oslo_vmware.api [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390331, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059611} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.920654] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1045.921421] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-019a0f10-68f7-4d28-bd27-8bd6f04aac86 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.943910] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] 58c62ac1-203c-4370-8c94-420d655b7c08/58c62ac1-203c-4370-8c94-420d655b7c08.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1045.944272] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-22c89d2e-2126-48de-a976-2a1b4ab8f25a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.964902] env[62965]: DEBUG oslo_vmware.api [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 1045.964902] env[62965]: value = "task-1390332" [ 1045.964902] env[62965]: _type = "Task" [ 1045.964902] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.973998] env[62965]: DEBUG oslo_vmware.api [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390332, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.005552] env[62965]: DEBUG nova.compute.manager [req-de0e51ac-e395-4cf6-af1b-5000a79fae3b req-556436be-a3d4-4ca8-b6a4-eaee48711230 service nova] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Received event network-vif-plugged-5f939a36-bc5f-4fd7-b29e-011f72b3256c {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1046.005830] env[62965]: DEBUG oslo_concurrency.lockutils [req-de0e51ac-e395-4cf6-af1b-5000a79fae3b req-556436be-a3d4-4ca8-b6a4-eaee48711230 service nova] Acquiring lock "73a7ca0e-b1d3-4464-969b-5111d2506a8b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.006131] env[62965]: DEBUG oslo_concurrency.lockutils [req-de0e51ac-e395-4cf6-af1b-5000a79fae3b req-556436be-a3d4-4ca8-b6a4-eaee48711230 service nova] Lock "73a7ca0e-b1d3-4464-969b-5111d2506a8b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.006306] env[62965]: DEBUG oslo_concurrency.lockutils [req-de0e51ac-e395-4cf6-af1b-5000a79fae3b req-556436be-a3d4-4ca8-b6a4-eaee48711230 service nova] Lock "73a7ca0e-b1d3-4464-969b-5111d2506a8b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.006503] env[62965]: DEBUG nova.compute.manager [req-de0e51ac-e395-4cf6-af1b-5000a79fae3b req-556436be-a3d4-4ca8-b6a4-eaee48711230 service nova] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] No waiting events found dispatching network-vif-plugged-5f939a36-bc5f-4fd7-b29e-011f72b3256c {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1046.006693] env[62965]: WARNING nova.compute.manager [req-de0e51ac-e395-4cf6-af1b-5000a79fae3b req-556436be-a3d4-4ca8-b6a4-eaee48711230 service nova] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Received unexpected event network-vif-plugged-5f939a36-bc5f-4fd7-b29e-011f72b3256c for instance with vm_state building and task_state spawning. [ 1046.007704] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.196s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.008222] env[62965]: DEBUG nova.compute.manager [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1046.010856] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4fa857a1-4a4c-4f0d-8ac0-6abcd392cc99 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.412s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.011125] env[62965]: DEBUG nova.objects.instance [None req-4fa857a1-4a4c-4f0d-8ac0-6abcd392cc99 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lazy-loading 'resources' on Instance uuid df03e327-959c-42af-85cd-a7af2ec4d0f7 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1046.095413] env[62965]: DEBUG nova.network.neutron [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Successfully updated port: 5f939a36-bc5f-4fd7-b29e-011f72b3256c {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1046.126891] env[62965]: DEBUG oslo_concurrency.lockutils [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "refresh_cache-c2448b76-1553-4b68-a731-0bb0967f4c1d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.127110] env[62965]: DEBUG oslo_concurrency.lockutils [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquired lock "refresh_cache-c2448b76-1553-4b68-a731-0bb0967f4c1d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.127297] env[62965]: DEBUG nova.network.neutron [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1046.474945] env[62965]: DEBUG oslo_vmware.api [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390332, 'name': ReconfigVM_Task, 'duration_secs': 0.268483} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.475208] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Reconfigured VM instance instance-00000065 to attach disk [datastore2] 58c62ac1-203c-4370-8c94-420d655b7c08/58c62ac1-203c-4370-8c94-420d655b7c08.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1046.475835] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0f4db79b-45b9-4cb8-815a-237e7f1d6440 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.481833] env[62965]: DEBUG oslo_vmware.api [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 1046.481833] env[62965]: value = "task-1390333" [ 1046.481833] env[62965]: _type = "Task" [ 1046.481833] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.489288] env[62965]: DEBUG oslo_vmware.api [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390333, 'name': Rename_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.516132] env[62965]: DEBUG nova.compute.utils [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1046.517684] env[62965]: DEBUG nova.compute.manager [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1046.517842] env[62965]: DEBUG nova.network.neutron [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1046.598673] env[62965]: DEBUG oslo_concurrency.lockutils [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquiring lock "refresh_cache-73a7ca0e-b1d3-4464-969b-5111d2506a8b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.598831] env[62965]: DEBUG oslo_concurrency.lockutils [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquired lock "refresh_cache-73a7ca0e-b1d3-4464-969b-5111d2506a8b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.598980] env[62965]: DEBUG nova.network.neutron [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1046.626331] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-946824b8-6fd3-4568-b578-ce86f897b1cf {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.635600] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5559991-293c-43d1-a930-1a6a7b48a433 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.669333] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-383c5251-a919-435f-b7ab-840cb625fa67 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.676999] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63c0f949-a219-4387-a8d1-753278289805 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.690064] env[62965]: DEBUG nova.compute.provider_tree [None req-4fa857a1-4a4c-4f0d-8ac0-6abcd392cc99 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1046.732437] env[62965]: DEBUG nova.policy [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03c0a9a0ec3c46bf883c68d86cb816d6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '978666785b0743bb843de96585f96fb4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 1046.860475] env[62965]: DEBUG nova.network.neutron [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Updating instance_info_cache with network_info: [{"id": "57672e8b-1df1-463d-b9b3-8290667d9fd2", "address": "fa:16:3e:53:46:47", "network": {"id": "4649fed7-c2ca-43a9-a183-7c99423da726", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1978698173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fab418e2fe1420793517663574b43bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43ad01d2-c7dd-453c-a929-8ad76294d13c", "external-id": "nsx-vlan-transportzone-176", "segmentation_id": 176, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57672e8b-1d", "ovs_interfaceid": "57672e8b-1df1-463d-b9b3-8290667d9fd2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.962437] env[62965]: DEBUG nova.network.neutron [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Successfully created port: 52432dbd-16ac-41f2-8bee-6d27c80125dd {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1046.991394] env[62965]: DEBUG oslo_vmware.api [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390333, 'name': Rename_Task, 'duration_secs': 0.131363} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.991633] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1046.991872] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aa27c9f6-a932-47bd-bbd8-7383e6a198ee {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.997844] env[62965]: DEBUG oslo_vmware.api [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 1046.997844] env[62965]: value = "task-1390334" [ 1046.997844] env[62965]: _type = "Task" [ 1046.997844] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.007089] env[62965]: DEBUG oslo_vmware.api [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390334, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.020692] env[62965]: DEBUG nova.compute.manager [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1047.131619] env[62965]: DEBUG nova.network.neutron [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1047.195422] env[62965]: DEBUG nova.scheduler.client.report [None req-4fa857a1-4a4c-4f0d-8ac0-6abcd392cc99 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1047.264841] env[62965]: DEBUG nova.network.neutron [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Updating instance_info_cache with network_info: [{"id": "5f939a36-bc5f-4fd7-b29e-011f72b3256c", "address": "fa:16:3e:e3:06:99", "network": {"id": "49dd42b0-6397-4fdd-a319-ed920c72a5af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-373172566-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d91c99b14018469e8998fd1ccc0fab4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f939a36-bc", "ovs_interfaceid": "5f939a36-bc5f-4fd7-b29e-011f72b3256c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.363311] env[62965]: DEBUG oslo_concurrency.lockutils [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Releasing lock "refresh_cache-c2448b76-1553-4b68-a731-0bb0967f4c1d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.508166] env[62965]: DEBUG oslo_vmware.api [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390334, 'name': PowerOnVM_Task, 'duration_secs': 0.410864} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.508528] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1047.508788] env[62965]: INFO nova.compute.manager [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Took 7.28 seconds to spawn the instance on the hypervisor. [ 1047.509014] env[62965]: DEBUG nova.compute.manager [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1047.509807] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f510d0d-93cb-4718-97a9-d1f3d50d1cbb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.700151] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4fa857a1-4a4c-4f0d-8ac0-6abcd392cc99 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.689s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.718700] env[62965]: INFO nova.scheduler.client.report [None req-4fa857a1-4a4c-4f0d-8ac0-6abcd392cc99 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Deleted allocations for instance df03e327-959c-42af-85cd-a7af2ec4d0f7 [ 1047.768549] env[62965]: DEBUG oslo_concurrency.lockutils [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Releasing lock "refresh_cache-73a7ca0e-b1d3-4464-969b-5111d2506a8b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.768549] env[62965]: DEBUG nova.compute.manager [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Instance network_info: |[{"id": "5f939a36-bc5f-4fd7-b29e-011f72b3256c", "address": "fa:16:3e:e3:06:99", "network": {"id": "49dd42b0-6397-4fdd-a319-ed920c72a5af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-373172566-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d91c99b14018469e8998fd1ccc0fab4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f939a36-bc", "ovs_interfaceid": "5f939a36-bc5f-4fd7-b29e-011f72b3256c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1047.768549] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e3:06:99', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f77ff7a1-209c-4f3f-b2a0-fd817741e739', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5f939a36-bc5f-4fd7-b29e-011f72b3256c', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1047.775310] env[62965]: DEBUG oslo.service.loopingcall [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1047.775733] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1047.775968] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6d07e739-ea8e-4139-a73e-38297835eac2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.796392] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1047.796392] env[62965]: value = "task-1390335" [ 1047.796392] env[62965]: _type = "Task" [ 1047.796392] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.803852] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390335, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.885680] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4edef041-5a69-475c-868a-700326298b10 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.905075] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34fa6735-0ea0-47ec-8efb-55e210d0496b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.912107] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Updating instance 'c2448b76-1553-4b68-a731-0bb0967f4c1d' progress to 83 {{(pid=62965) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1048.029226] env[62965]: INFO nova.compute.manager [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Took 16.50 seconds to build instance. [ 1048.030475] env[62965]: DEBUG nova.compute.manager [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1048.035442] env[62965]: DEBUG nova.compute.manager [req-b145df4b-f274-4f66-91ce-cea6fb3218d5 req-108562df-7170-4b52-a263-1c04a7cdb6cd service nova] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Received event network-changed-5f939a36-bc5f-4fd7-b29e-011f72b3256c {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1048.035442] env[62965]: DEBUG nova.compute.manager [req-b145df4b-f274-4f66-91ce-cea6fb3218d5 req-108562df-7170-4b52-a263-1c04a7cdb6cd service nova] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Refreshing instance network info cache due to event network-changed-5f939a36-bc5f-4fd7-b29e-011f72b3256c. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 1048.035442] env[62965]: DEBUG oslo_concurrency.lockutils [req-b145df4b-f274-4f66-91ce-cea6fb3218d5 req-108562df-7170-4b52-a263-1c04a7cdb6cd service nova] Acquiring lock "refresh_cache-73a7ca0e-b1d3-4464-969b-5111d2506a8b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.035442] env[62965]: DEBUG oslo_concurrency.lockutils [req-b145df4b-f274-4f66-91ce-cea6fb3218d5 req-108562df-7170-4b52-a263-1c04a7cdb6cd service nova] Acquired lock "refresh_cache-73a7ca0e-b1d3-4464-969b-5111d2506a8b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.035442] env[62965]: DEBUG nova.network.neutron [req-b145df4b-f274-4f66-91ce-cea6fb3218d5 req-108562df-7170-4b52-a263-1c04a7cdb6cd service nova] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Refreshing network info cache for port 5f939a36-bc5f-4fd7-b29e-011f72b3256c {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1048.056481] env[62965]: DEBUG nova.virt.hardware [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1048.056727] env[62965]: DEBUG nova.virt.hardware [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1048.056889] env[62965]: DEBUG nova.virt.hardware [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1048.057089] env[62965]: DEBUG nova.virt.hardware [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1048.057298] env[62965]: DEBUG nova.virt.hardware [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1048.057393] env[62965]: DEBUG nova.virt.hardware [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1048.057654] env[62965]: DEBUG nova.virt.hardware [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1048.057831] env[62965]: DEBUG nova.virt.hardware [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1048.058481] env[62965]: DEBUG nova.virt.hardware [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1048.058481] env[62965]: DEBUG nova.virt.hardware [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1048.058481] env[62965]: DEBUG nova.virt.hardware [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1048.059629] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a33200d9-e78c-425e-803d-27847d475f28 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.067726] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05b06573-1f3d-4e5d-9ad1-c81ea8fb5eea {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.226825] env[62965]: DEBUG oslo_concurrency.lockutils [None req-4fa857a1-4a4c-4f0d-8ac0-6abcd392cc99 tempest-ServerDiskConfigTestJSON-370570873 tempest-ServerDiskConfigTestJSON-370570873-project-member] Lock "df03e327-959c-42af-85cd-a7af2ec4d0f7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.983s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.307070] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390335, 'name': CreateVM_Task, 'duration_secs': 0.286338} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.307252] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1048.307946] env[62965]: DEBUG oslo_concurrency.lockutils [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.308133] env[62965]: DEBUG oslo_concurrency.lockutils [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.308476] env[62965]: DEBUG oslo_concurrency.lockutils [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1048.308752] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8febe960-4304-44c0-85f1-f50d7c11a0bb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.313279] env[62965]: DEBUG oslo_vmware.api [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 1048.313279] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52911864-142c-ef9b-8f60-c7a10ba689f0" [ 1048.313279] env[62965]: _type = "Task" [ 1048.313279] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.322482] env[62965]: DEBUG oslo_vmware.api [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52911864-142c-ef9b-8f60-c7a10ba689f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.345281] env[62965]: DEBUG nova.compute.manager [req-b8a80f76-c170-4253-a4ea-d5769587e05a req-c00b90ce-d6c7-4167-9863-77eba0b65a2b service nova] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Received event network-vif-plugged-52432dbd-16ac-41f2-8bee-6d27c80125dd {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1048.345939] env[62965]: DEBUG oslo_concurrency.lockutils [req-b8a80f76-c170-4253-a4ea-d5769587e05a req-c00b90ce-d6c7-4167-9863-77eba0b65a2b service nova] Acquiring lock "88e45c29-4d21-4f91-92f4-fdebaf905ccb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.345939] env[62965]: DEBUG oslo_concurrency.lockutils [req-b8a80f76-c170-4253-a4ea-d5769587e05a req-c00b90ce-d6c7-4167-9863-77eba0b65a2b service nova] Lock "88e45c29-4d21-4f91-92f4-fdebaf905ccb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.345939] env[62965]: DEBUG oslo_concurrency.lockutils [req-b8a80f76-c170-4253-a4ea-d5769587e05a req-c00b90ce-d6c7-4167-9863-77eba0b65a2b service nova] Lock "88e45c29-4d21-4f91-92f4-fdebaf905ccb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.346083] env[62965]: DEBUG nova.compute.manager [req-b8a80f76-c170-4253-a4ea-d5769587e05a req-c00b90ce-d6c7-4167-9863-77eba0b65a2b service nova] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] No waiting events found dispatching network-vif-plugged-52432dbd-16ac-41f2-8bee-6d27c80125dd {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1048.346254] env[62965]: WARNING nova.compute.manager [req-b8a80f76-c170-4253-a4ea-d5769587e05a req-c00b90ce-d6c7-4167-9863-77eba0b65a2b service nova] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Received unexpected event network-vif-plugged-52432dbd-16ac-41f2-8bee-6d27c80125dd for instance with vm_state building and task_state spawning. [ 1048.418560] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1048.418959] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8ef26c39-efdb-4280-85f9-003af0d3eded {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.426171] env[62965]: DEBUG oslo_vmware.api [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1048.426171] env[62965]: value = "task-1390336" [ 1048.426171] env[62965]: _type = "Task" [ 1048.426171] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.430185] env[62965]: DEBUG nova.network.neutron [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Successfully updated port: 52432dbd-16ac-41f2-8bee-6d27c80125dd {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1048.436891] env[62965]: DEBUG oslo_vmware.api [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390336, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.533925] env[62965]: DEBUG oslo_concurrency.lockutils [None req-13c62fba-61cd-409b-8e91-f0c36d9a8db2 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "58c62ac1-203c-4370-8c94-420d655b7c08" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.008s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.751017] env[62965]: DEBUG nova.network.neutron [req-b145df4b-f274-4f66-91ce-cea6fb3218d5 req-108562df-7170-4b52-a263-1c04a7cdb6cd service nova] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Updated VIF entry in instance network info cache for port 5f939a36-bc5f-4fd7-b29e-011f72b3256c. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1048.751515] env[62965]: DEBUG nova.network.neutron [req-b145df4b-f274-4f66-91ce-cea6fb3218d5 req-108562df-7170-4b52-a263-1c04a7cdb6cd service nova] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Updating instance_info_cache with network_info: [{"id": "5f939a36-bc5f-4fd7-b29e-011f72b3256c", "address": "fa:16:3e:e3:06:99", "network": {"id": "49dd42b0-6397-4fdd-a319-ed920c72a5af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-373172566-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d91c99b14018469e8998fd1ccc0fab4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f939a36-bc", "ovs_interfaceid": "5f939a36-bc5f-4fd7-b29e-011f72b3256c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1048.823623] env[62965]: DEBUG oslo_vmware.api [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52911864-142c-ef9b-8f60-c7a10ba689f0, 'name': SearchDatastore_Task, 'duration_secs': 0.009263} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.823872] env[62965]: DEBUG oslo_concurrency.lockutils [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.824119] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1048.824360] env[62965]: DEBUG oslo_concurrency.lockutils [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.824513] env[62965]: DEBUG oslo_concurrency.lockutils [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.824695] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1048.824960] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0dd63d4d-3b88-4657-b168-33cdeefb1e98 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.832582] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1048.832760] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1048.833456] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc85c7ab-0dad-4c81-b9e4-ddf74dafe8f5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.838390] env[62965]: DEBUG oslo_vmware.api [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 1048.838390] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52f884f4-25d6-ec75-c122-cb3c1a408a5d" [ 1048.838390] env[62965]: _type = "Task" [ 1048.838390] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.845509] env[62965]: DEBUG oslo_vmware.api [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52f884f4-25d6-ec75-c122-cb3c1a408a5d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.938089] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "refresh_cache-88e45c29-4d21-4f91-92f4-fdebaf905ccb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.938338] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquired lock "refresh_cache-88e45c29-4d21-4f91-92f4-fdebaf905ccb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.938444] env[62965]: DEBUG nova.network.neutron [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1048.939664] env[62965]: DEBUG oslo_vmware.api [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390336, 'name': PowerOnVM_Task, 'duration_secs': 0.361931} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.940143] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1048.940395] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-58c3aa20-cfbd-4ac9-9f2a-1194451dc957 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Updating instance 'c2448b76-1553-4b68-a731-0bb0967f4c1d' progress to 100 {{(pid=62965) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1049.254741] env[62965]: DEBUG oslo_concurrency.lockutils [req-b145df4b-f274-4f66-91ce-cea6fb3218d5 req-108562df-7170-4b52-a263-1c04a7cdb6cd service nova] Releasing lock "refresh_cache-73a7ca0e-b1d3-4464-969b-5111d2506a8b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.349143] env[62965]: DEBUG oslo_vmware.api [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52f884f4-25d6-ec75-c122-cb3c1a408a5d, 'name': SearchDatastore_Task, 'duration_secs': 0.008354} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.350024] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e3d191a9-7e40-4f77-8d95-9bdd3ebe099e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.355222] env[62965]: DEBUG oslo_vmware.api [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 1049.355222] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52a5d513-0ead-7d38-f1f0-abe998ab5bdc" [ 1049.355222] env[62965]: _type = "Task" [ 1049.355222] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.362824] env[62965]: DEBUG oslo_vmware.api [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52a5d513-0ead-7d38-f1f0-abe998ab5bdc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.473021] env[62965]: DEBUG nova.network.neutron [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1049.540490] env[62965]: DEBUG oslo_concurrency.lockutils [None req-fc31726e-6445-4e1c-b8f7-59f18666630b tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "58c62ac1-203c-4370-8c94-420d655b7c08" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.540750] env[62965]: DEBUG oslo_concurrency.lockutils [None req-fc31726e-6445-4e1c-b8f7-59f18666630b tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "58c62ac1-203c-4370-8c94-420d655b7c08" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.540930] env[62965]: DEBUG nova.compute.manager [None req-fc31726e-6445-4e1c-b8f7-59f18666630b tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1049.541903] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-305236b1-99e7-41e3-a255-1106705411fd {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.548769] env[62965]: DEBUG nova.compute.manager [None req-fc31726e-6445-4e1c-b8f7-59f18666630b tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62965) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 1049.549323] env[62965]: DEBUG nova.objects.instance [None req-fc31726e-6445-4e1c-b8f7-59f18666630b tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lazy-loading 'flavor' on Instance uuid 58c62ac1-203c-4370-8c94-420d655b7c08 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1049.600794] env[62965]: DEBUG nova.network.neutron [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Updating instance_info_cache with network_info: [{"id": "52432dbd-16ac-41f2-8bee-6d27c80125dd", "address": "fa:16:3e:0e:97:c7", "network": {"id": "62052700-1781-4933-93c8-fa1aafdb0fd2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1931798228-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "978666785b0743bb843de96585f96fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52432dbd-16", "ovs_interfaceid": "52432dbd-16ac-41f2-8bee-6d27c80125dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.865543] env[62965]: DEBUG oslo_vmware.api [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52a5d513-0ead-7d38-f1f0-abe998ab5bdc, 'name': SearchDatastore_Task, 'duration_secs': 0.009506} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.865911] env[62965]: DEBUG oslo_concurrency.lockutils [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.866268] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 73a7ca0e-b1d3-4464-969b-5111d2506a8b/73a7ca0e-b1d3-4464-969b-5111d2506a8b.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1049.866590] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9008daf9-09be-4dd2-8eff-65a0c3d215cc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.873913] env[62965]: DEBUG oslo_vmware.api [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 1049.873913] env[62965]: value = "task-1390337" [ 1049.873913] env[62965]: _type = "Task" [ 1049.873913] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.882110] env[62965]: DEBUG oslo_vmware.api [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390337, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.104372] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Releasing lock "refresh_cache-88e45c29-4d21-4f91-92f4-fdebaf905ccb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1050.104669] env[62965]: DEBUG nova.compute.manager [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Instance network_info: |[{"id": "52432dbd-16ac-41f2-8bee-6d27c80125dd", "address": "fa:16:3e:0e:97:c7", "network": {"id": "62052700-1781-4933-93c8-fa1aafdb0fd2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1931798228-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "978666785b0743bb843de96585f96fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52432dbd-16", "ovs_interfaceid": "52432dbd-16ac-41f2-8bee-6d27c80125dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1050.105565] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0e:97:c7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a10c88d7-d13f-44fd-acee-7a734eb5f56a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '52432dbd-16ac-41f2-8bee-6d27c80125dd', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1050.114339] env[62965]: DEBUG oslo.service.loopingcall [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1050.114639] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1050.115395] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-23ac482a-082d-4b4f-8625-93067bb08030 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.139861] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1050.139861] env[62965]: value = "task-1390338" [ 1050.139861] env[62965]: _type = "Task" [ 1050.139861] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.150023] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390338, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.373063] env[62965]: DEBUG nova.compute.manager [req-efa87cb9-e31e-448a-bd2d-1451467f101a req-31a4d818-173b-4494-8916-3be4dce97b6e service nova] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Received event network-changed-52432dbd-16ac-41f2-8bee-6d27c80125dd {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1050.373277] env[62965]: DEBUG nova.compute.manager [req-efa87cb9-e31e-448a-bd2d-1451467f101a req-31a4d818-173b-4494-8916-3be4dce97b6e service nova] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Refreshing instance network info cache due to event network-changed-52432dbd-16ac-41f2-8bee-6d27c80125dd. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 1050.373509] env[62965]: DEBUG oslo_concurrency.lockutils [req-efa87cb9-e31e-448a-bd2d-1451467f101a req-31a4d818-173b-4494-8916-3be4dce97b6e service nova] Acquiring lock "refresh_cache-88e45c29-4d21-4f91-92f4-fdebaf905ccb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1050.373842] env[62965]: DEBUG oslo_concurrency.lockutils [req-efa87cb9-e31e-448a-bd2d-1451467f101a req-31a4d818-173b-4494-8916-3be4dce97b6e service nova] Acquired lock "refresh_cache-88e45c29-4d21-4f91-92f4-fdebaf905ccb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.373842] env[62965]: DEBUG nova.network.neutron [req-efa87cb9-e31e-448a-bd2d-1451467f101a req-31a4d818-173b-4494-8916-3be4dce97b6e service nova] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Refreshing network info cache for port 52432dbd-16ac-41f2-8bee-6d27c80125dd {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1050.384327] env[62965]: DEBUG oslo_vmware.api [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390337, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.559761] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc31726e-6445-4e1c-b8f7-59f18666630b tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1050.560125] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d3ef5626-8fc8-442d-8937-364027af3db1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.567993] env[62965]: DEBUG oslo_vmware.api [None req-fc31726e-6445-4e1c-b8f7-59f18666630b tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 1050.567993] env[62965]: value = "task-1390339" [ 1050.567993] env[62965]: _type = "Task" [ 1050.567993] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.575151] env[62965]: DEBUG oslo_vmware.api [None req-fc31726e-6445-4e1c-b8f7-59f18666630b tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390339, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.649933] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390338, 'name': CreateVM_Task, 'duration_secs': 0.479701} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.650125] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1050.650849] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1050.651050] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1050.651431] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1050.651683] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-43df97ed-83e5-4832-b076-2912656c633a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.656083] env[62965]: DEBUG oslo_vmware.api [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 1050.656083] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5211b3b2-298c-6e2f-42d8-6d83f0366859" [ 1050.656083] env[62965]: _type = "Task" [ 1050.656083] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.664257] env[62965]: DEBUG oslo_vmware.api [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5211b3b2-298c-6e2f-42d8-6d83f0366859, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.889324] env[62965]: DEBUG oslo_vmware.api [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390337, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.524636} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.890700] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 73a7ca0e-b1d3-4464-969b-5111d2506a8b/73a7ca0e-b1d3-4464-969b-5111d2506a8b.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1050.890700] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1050.890700] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a270be98-091e-4d6c-9748-8fe3b890cb24 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.897607] env[62965]: DEBUG oslo_vmware.api [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 1050.897607] env[62965]: value = "task-1390340" [ 1050.897607] env[62965]: _type = "Task" [ 1050.897607] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.907716] env[62965]: DEBUG oslo_vmware.api [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390340, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.082893] env[62965]: DEBUG oslo_vmware.api [None req-fc31726e-6445-4e1c-b8f7-59f18666630b tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390339, 'name': PowerOffVM_Task, 'duration_secs': 0.169897} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.082893] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc31726e-6445-4e1c-b8f7-59f18666630b tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1051.082893] env[62965]: DEBUG nova.compute.manager [None req-fc31726e-6445-4e1c-b8f7-59f18666630b tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1051.082893] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de6cca8-4a3c-4873-adc1-e17d683504b5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.167801] env[62965]: DEBUG oslo_vmware.api [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5211b3b2-298c-6e2f-42d8-6d83f0366859, 'name': SearchDatastore_Task, 'duration_secs': 0.008265} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.168159] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.168406] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1051.168658] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1051.168940] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1051.169356] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1051.169613] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4e8f210c-5957-4bfc-b36f-cac2e9deced9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.177729] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1051.177918] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1051.178631] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2303eebd-0496-4264-85c7-ff46d5684dca {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.186260] env[62965]: DEBUG oslo_vmware.api [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 1051.186260] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52d069bd-0f78-a06a-b155-5ef916300a04" [ 1051.186260] env[62965]: _type = "Task" [ 1051.186260] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.194821] env[62965]: DEBUG oslo_vmware.api [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52d069bd-0f78-a06a-b155-5ef916300a04, 'name': SearchDatastore_Task, 'duration_secs': 0.008297} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.195443] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f7a07bf1-01f3-421b-ade4-0daa52ed7a9d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.201299] env[62965]: DEBUG oslo_vmware.api [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 1051.201299] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52eb4480-92fb-7f90-f53b-3d7a07dbf123" [ 1051.201299] env[62965]: _type = "Task" [ 1051.201299] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.208907] env[62965]: DEBUG oslo_vmware.api [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52eb4480-92fb-7f90-f53b-3d7a07dbf123, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.219713] env[62965]: DEBUG nova.network.neutron [req-efa87cb9-e31e-448a-bd2d-1451467f101a req-31a4d818-173b-4494-8916-3be4dce97b6e service nova] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Updated VIF entry in instance network info cache for port 52432dbd-16ac-41f2-8bee-6d27c80125dd. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1051.220078] env[62965]: DEBUG nova.network.neutron [req-efa87cb9-e31e-448a-bd2d-1451467f101a req-31a4d818-173b-4494-8916-3be4dce97b6e service nova] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Updating instance_info_cache with network_info: [{"id": "52432dbd-16ac-41f2-8bee-6d27c80125dd", "address": "fa:16:3e:0e:97:c7", "network": {"id": "62052700-1781-4933-93c8-fa1aafdb0fd2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1931798228-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "978666785b0743bb843de96585f96fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52432dbd-16", "ovs_interfaceid": "52432dbd-16ac-41f2-8bee-6d27c80125dd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.407038] env[62965]: DEBUG oslo_vmware.api [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390340, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080088} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.407336] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1051.408142] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f83d175-b313-4a6f-8522-35c41b8eb7f7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.430559] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 73a7ca0e-b1d3-4464-969b-5111d2506a8b/73a7ca0e-b1d3-4464-969b-5111d2506a8b.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1051.430942] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-293a3302-7f13-4a90-b949-6916adc71973 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.452258] env[62965]: DEBUG oslo_vmware.api [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 1051.452258] env[62965]: value = "task-1390341" [ 1051.452258] env[62965]: _type = "Task" [ 1051.452258] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.460806] env[62965]: DEBUG oslo_vmware.api [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390341, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.546298] env[62965]: DEBUG oslo_concurrency.lockutils [None req-eac9b66d-1384-4e35-8f5d-c3800838412a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "c2448b76-1553-4b68-a731-0bb0967f4c1d" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.546582] env[62965]: DEBUG oslo_concurrency.lockutils [None req-eac9b66d-1384-4e35-8f5d-c3800838412a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "c2448b76-1553-4b68-a731-0bb0967f4c1d" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.546771] env[62965]: DEBUG nova.compute.manager [None req-eac9b66d-1384-4e35-8f5d-c3800838412a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Going to confirm migration 3 {{(pid=62965) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5112}} [ 1051.594998] env[62965]: DEBUG oslo_concurrency.lockutils [None req-fc31726e-6445-4e1c-b8f7-59f18666630b tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "58c62ac1-203c-4370-8c94-420d655b7c08" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.054s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.713699] env[62965]: DEBUG oslo_vmware.api [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52eb4480-92fb-7f90-f53b-3d7a07dbf123, 'name': SearchDatastore_Task, 'duration_secs': 0.009067} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.714044] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.714357] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 88e45c29-4d21-4f91-92f4-fdebaf905ccb/88e45c29-4d21-4f91-92f4-fdebaf905ccb.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1051.714652] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ebe532bb-a968-4615-9924-870fdb7fc275 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.721380] env[62965]: DEBUG oslo_vmware.api [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 1051.721380] env[62965]: value = "task-1390342" [ 1051.721380] env[62965]: _type = "Task" [ 1051.721380] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.721971] env[62965]: DEBUG oslo_concurrency.lockutils [req-efa87cb9-e31e-448a-bd2d-1451467f101a req-31a4d818-173b-4494-8916-3be4dce97b6e service nova] Releasing lock "refresh_cache-88e45c29-4d21-4f91-92f4-fdebaf905ccb" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1051.729668] env[62965]: DEBUG oslo_vmware.api [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390342, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.964649] env[62965]: DEBUG oslo_vmware.api [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390341, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.133486] env[62965]: DEBUG oslo_concurrency.lockutils [None req-eac9b66d-1384-4e35-8f5d-c3800838412a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "refresh_cache-c2448b76-1553-4b68-a731-0bb0967f4c1d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1052.133808] env[62965]: DEBUG oslo_concurrency.lockutils [None req-eac9b66d-1384-4e35-8f5d-c3800838412a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquired lock "refresh_cache-c2448b76-1553-4b68-a731-0bb0967f4c1d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1052.133919] env[62965]: DEBUG nova.network.neutron [None req-eac9b66d-1384-4e35-8f5d-c3800838412a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1052.134018] env[62965]: DEBUG nova.objects.instance [None req-eac9b66d-1384-4e35-8f5d-c3800838412a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lazy-loading 'info_cache' on Instance uuid c2448b76-1553-4b68-a731-0bb0967f4c1d {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1052.233877] env[62965]: DEBUG oslo_vmware.api [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390342, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.333220] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f91cb700-c247-47d4-a041-bf4e04bfb125 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "58c62ac1-203c-4370-8c94-420d655b7c08" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.334508] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f91cb700-c247-47d4-a041-bf4e04bfb125 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "58c62ac1-203c-4370-8c94-420d655b7c08" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.334508] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f91cb700-c247-47d4-a041-bf4e04bfb125 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "58c62ac1-203c-4370-8c94-420d655b7c08-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.334508] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f91cb700-c247-47d4-a041-bf4e04bfb125 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "58c62ac1-203c-4370-8c94-420d655b7c08-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.334508] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f91cb700-c247-47d4-a041-bf4e04bfb125 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "58c62ac1-203c-4370-8c94-420d655b7c08-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.336732] env[62965]: INFO nova.compute.manager [None req-f91cb700-c247-47d4-a041-bf4e04bfb125 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Terminating instance [ 1052.463619] env[62965]: DEBUG oslo_vmware.api [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390341, 'name': ReconfigVM_Task, 'duration_secs': 0.596945} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.463946] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 73a7ca0e-b1d3-4464-969b-5111d2506a8b/73a7ca0e-b1d3-4464-969b-5111d2506a8b.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1052.464612] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9540744b-798b-4077-84d5-d245ca8cc5ee {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.472098] env[62965]: DEBUG oslo_vmware.api [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 1052.472098] env[62965]: value = "task-1390343" [ 1052.472098] env[62965]: _type = "Task" [ 1052.472098] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.481656] env[62965]: DEBUG oslo_vmware.api [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390343, 'name': Rename_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.734793] env[62965]: DEBUG oslo_vmware.api [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390342, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.539556} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.735032] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 88e45c29-4d21-4f91-92f4-fdebaf905ccb/88e45c29-4d21-4f91-92f4-fdebaf905ccb.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1052.735238] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1052.735497] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a87469c0-37d2-4da3-b350-06e4030ec198 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.742842] env[62965]: DEBUG oslo_vmware.api [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 1052.742842] env[62965]: value = "task-1390344" [ 1052.742842] env[62965]: _type = "Task" [ 1052.742842] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.750770] env[62965]: DEBUG oslo_vmware.api [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390344, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.840318] env[62965]: DEBUG nova.compute.manager [None req-f91cb700-c247-47d4-a041-bf4e04bfb125 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1052.840586] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-f91cb700-c247-47d4-a041-bf4e04bfb125 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1052.843913] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fb831ba-b9f1-4db1-b833-f192417fce16 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.849349] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-f91cb700-c247-47d4-a041-bf4e04bfb125 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1052.849594] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3904fddd-7034-49c3-8c49-c7489971a087 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.982934] env[62965]: DEBUG oslo_vmware.api [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390343, 'name': Rename_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.062924] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-f91cb700-c247-47d4-a041-bf4e04bfb125 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1053.063103] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-f91cb700-c247-47d4-a041-bf4e04bfb125 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Deleting contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1053.063336] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-f91cb700-c247-47d4-a041-bf4e04bfb125 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Deleting the datastore file [datastore2] 58c62ac1-203c-4370-8c94-420d655b7c08 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1053.063579] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5296b78f-4208-42fc-8a27-523f62798954 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.070858] env[62965]: DEBUG oslo_vmware.api [None req-f91cb700-c247-47d4-a041-bf4e04bfb125 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 1053.070858] env[62965]: value = "task-1390346" [ 1053.070858] env[62965]: _type = "Task" [ 1053.070858] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.080443] env[62965]: DEBUG oslo_vmware.api [None req-f91cb700-c247-47d4-a041-bf4e04bfb125 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390346, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.255661] env[62965]: DEBUG oslo_vmware.api [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390344, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070109} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.255994] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1053.256827] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d56dae55-30a8-49e9-a6ec-dc53152406f6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.278964] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] 88e45c29-4d21-4f91-92f4-fdebaf905ccb/88e45c29-4d21-4f91-92f4-fdebaf905ccb.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1053.282516] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7e9ac2e7-d0a0-4901-b22e-2e140039177b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.303099] env[62965]: DEBUG oslo_vmware.api [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 1053.303099] env[62965]: value = "task-1390347" [ 1053.303099] env[62965]: _type = "Task" [ 1053.303099] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.311145] env[62965]: DEBUG oslo_vmware.api [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390347, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.403638] env[62965]: DEBUG nova.network.neutron [None req-eac9b66d-1384-4e35-8f5d-c3800838412a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Updating instance_info_cache with network_info: [{"id": "57672e8b-1df1-463d-b9b3-8290667d9fd2", "address": "fa:16:3e:53:46:47", "network": {"id": "4649fed7-c2ca-43a9-a183-7c99423da726", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1978698173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fab418e2fe1420793517663574b43bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43ad01d2-c7dd-453c-a929-8ad76294d13c", "external-id": "nsx-vlan-transportzone-176", "segmentation_id": 176, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap57672e8b-1d", "ovs_interfaceid": "57672e8b-1df1-463d-b9b3-8290667d9fd2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.483171] env[62965]: DEBUG oslo_vmware.api [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390343, 'name': Rename_Task, 'duration_secs': 0.566924} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.483481] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1053.483738] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-46eac741-10e9-4512-ad99-8442cf814678 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.490748] env[62965]: DEBUG oslo_vmware.api [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 1053.490748] env[62965]: value = "task-1390348" [ 1053.490748] env[62965]: _type = "Task" [ 1053.490748] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.502631] env[62965]: DEBUG oslo_vmware.api [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390348, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.579780] env[62965]: DEBUG oslo_vmware.api [None req-f91cb700-c247-47d4-a041-bf4e04bfb125 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390346, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.38081} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.579998] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-f91cb700-c247-47d4-a041-bf4e04bfb125 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1053.580207] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-f91cb700-c247-47d4-a041-bf4e04bfb125 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Deleted contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1053.580405] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-f91cb700-c247-47d4-a041-bf4e04bfb125 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1053.580935] env[62965]: INFO nova.compute.manager [None req-f91cb700-c247-47d4-a041-bf4e04bfb125 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Took 0.74 seconds to destroy the instance on the hypervisor. [ 1053.581359] env[62965]: DEBUG oslo.service.loopingcall [None req-f91cb700-c247-47d4-a041-bf4e04bfb125 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1053.581602] env[62965]: DEBUG nova.compute.manager [-] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1053.581730] env[62965]: DEBUG nova.network.neutron [-] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1053.815363] env[62965]: DEBUG oslo_vmware.api [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390347, 'name': ReconfigVM_Task, 'duration_secs': 0.294041} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.815585] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Reconfigured VM instance instance-00000067 to attach disk [datastore2] 88e45c29-4d21-4f91-92f4-fdebaf905ccb/88e45c29-4d21-4f91-92f4-fdebaf905ccb.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1053.816227] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-74f27a19-737b-4e30-b163-119f50c766c7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.823780] env[62965]: DEBUG oslo_vmware.api [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 1053.823780] env[62965]: value = "task-1390349" [ 1053.823780] env[62965]: _type = "Task" [ 1053.823780] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.831664] env[62965]: DEBUG oslo_vmware.api [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390349, 'name': Rename_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.850802] env[62965]: DEBUG nova.compute.manager [req-897ceca5-b06d-4f66-b18f-4d2e0f4be87f req-f99b3dfa-43a7-4a19-8d98-c51ee22313e3 service nova] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Received event network-vif-deleted-d67556e3-a304-4a17-9d28-b275448ee50c {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1053.851075] env[62965]: INFO nova.compute.manager [req-897ceca5-b06d-4f66-b18f-4d2e0f4be87f req-f99b3dfa-43a7-4a19-8d98-c51ee22313e3 service nova] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Neutron deleted interface d67556e3-a304-4a17-9d28-b275448ee50c; detaching it from the instance and deleting it from the info cache [ 1053.851310] env[62965]: DEBUG nova.network.neutron [req-897ceca5-b06d-4f66-b18f-4d2e0f4be87f req-f99b3dfa-43a7-4a19-8d98-c51ee22313e3 service nova] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.905616] env[62965]: DEBUG oslo_concurrency.lockutils [None req-eac9b66d-1384-4e35-8f5d-c3800838412a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Releasing lock "refresh_cache-c2448b76-1553-4b68-a731-0bb0967f4c1d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1053.905899] env[62965]: DEBUG nova.objects.instance [None req-eac9b66d-1384-4e35-8f5d-c3800838412a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lazy-loading 'migration_context' on Instance uuid c2448b76-1553-4b68-a731-0bb0967f4c1d {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1054.000383] env[62965]: DEBUG oslo_vmware.api [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390348, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.327991] env[62965]: DEBUG nova.network.neutron [-] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.334676] env[62965]: DEBUG oslo_vmware.api [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390349, 'name': Rename_Task, 'duration_secs': 0.148895} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.334941] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1054.335207] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0f0956ba-e533-4adc-8855-c5b70b14e7bf {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.342601] env[62965]: DEBUG oslo_vmware.api [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 1054.342601] env[62965]: value = "task-1390350" [ 1054.342601] env[62965]: _type = "Task" [ 1054.342601] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.351715] env[62965]: DEBUG oslo_vmware.api [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390350, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.355366] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bdfa507f-3853-4360-a3ed-029672ea1479 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.363838] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f20c48cf-2f12-40e9-9913-8ee46e112206 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.394993] env[62965]: DEBUG nova.compute.manager [req-897ceca5-b06d-4f66-b18f-4d2e0f4be87f req-f99b3dfa-43a7-4a19-8d98-c51ee22313e3 service nova] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Detach interface failed, port_id=d67556e3-a304-4a17-9d28-b275448ee50c, reason: Instance 58c62ac1-203c-4370-8c94-420d655b7c08 could not be found. {{(pid=62965) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11294}} [ 1054.408155] env[62965]: DEBUG nova.objects.base [None req-eac9b66d-1384-4e35-8f5d-c3800838412a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62965) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1054.409136] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcda19f5-78b9-4786-8a4d-e42c75c00f04 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.430100] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a260e1e6-3729-41c7-b74d-d2873741ec89 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.437180] env[62965]: DEBUG oslo_vmware.api [None req-eac9b66d-1384-4e35-8f5d-c3800838412a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1054.437180] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52a6fc1e-1c52-71bf-7b72-d9c69e9f9bf2" [ 1054.437180] env[62965]: _type = "Task" [ 1054.437180] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.443439] env[62965]: DEBUG oslo_vmware.api [None req-eac9b66d-1384-4e35-8f5d-c3800838412a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52a6fc1e-1c52-71bf-7b72-d9c69e9f9bf2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.502077] env[62965]: DEBUG oslo_vmware.api [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390348, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.514477] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Acquiring lock "740b4d83-afc5-4ee2-9130-8c1d0697e5e5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.515461] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Lock "740b4d83-afc5-4ee2-9130-8c1d0697e5e5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.831554] env[62965]: INFO nova.compute.manager [-] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Took 1.25 seconds to deallocate network for instance. [ 1054.852399] env[62965]: DEBUG oslo_vmware.api [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390350, 'name': PowerOnVM_Task, 'duration_secs': 0.474286} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.852691] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1054.852911] env[62965]: INFO nova.compute.manager [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Took 6.82 seconds to spawn the instance on the hypervisor. [ 1054.853113] env[62965]: DEBUG nova.compute.manager [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1054.853950] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eb39dfc-37fe-41ad-96d4-4772a771a18e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.945372] env[62965]: DEBUG oslo_vmware.api [None req-eac9b66d-1384-4e35-8f5d-c3800838412a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52a6fc1e-1c52-71bf-7b72-d9c69e9f9bf2, 'name': SearchDatastore_Task, 'duration_secs': 0.006608} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.945715] env[62965]: DEBUG oslo_concurrency.lockutils [None req-eac9b66d-1384-4e35-8f5d-c3800838412a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.945978] env[62965]: DEBUG oslo_concurrency.lockutils [None req-eac9b66d-1384-4e35-8f5d-c3800838412a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.002604] env[62965]: DEBUG oslo_vmware.api [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390348, 'name': PowerOnVM_Task, 'duration_secs': 1.14254} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.002902] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1055.003127] env[62965]: INFO nova.compute.manager [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Took 9.17 seconds to spawn the instance on the hypervisor. [ 1055.003311] env[62965]: DEBUG nova.compute.manager [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1055.004069] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f60738d-942e-4443-9172-e263d1282a92 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.017876] env[62965]: DEBUG nova.compute.manager [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1055.337643] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f91cb700-c247-47d4-a041-bf4e04bfb125 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.371960] env[62965]: INFO nova.compute.manager [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Took 19.30 seconds to build instance. [ 1055.526491] env[62965]: INFO nova.compute.manager [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Took 20.15 seconds to build instance. [ 1055.540297] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.564095] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80579643-cf1a-4933-b3e9-c808a2dcc53d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.572667] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-042ca72f-bb01-482e-ba4d-07fa0a5c3403 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.604011] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3f40cfd-0e03-422a-8164-fa18ed8de00c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.611306] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27c03cdc-b180-4e8a-b6d2-6d515e68745c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.624804] env[62965]: DEBUG nova.compute.provider_tree [None req-eac9b66d-1384-4e35-8f5d-c3800838412a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1055.874394] env[62965]: DEBUG oslo_concurrency.lockutils [None req-20bff35f-63d4-40e5-be12-1ccfbd1040c7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "88e45c29-4d21-4f91-92f4-fdebaf905ccb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.812s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.028874] env[62965]: DEBUG oslo_concurrency.lockutils [None req-af857db8-4b3e-4ede-95a2-178d4224b4f2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "73a7ca0e-b1d3-4464-969b-5111d2506a8b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.658s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.065197] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d86a918e-a6ba-4395-ba95-0d0b87490af9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.071648] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-6ad61df8-aa1b-4430-888d-1428e935ccd7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Suspending the VM {{(pid=62965) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1056.071896] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-37d857b4-9b02-4d47-8d35-26cd56918aa3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.077895] env[62965]: DEBUG oslo_vmware.api [None req-6ad61df8-aa1b-4430-888d-1428e935ccd7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 1056.077895] env[62965]: value = "task-1390351" [ 1056.077895] env[62965]: _type = "Task" [ 1056.077895] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.088714] env[62965]: DEBUG oslo_vmware.api [None req-6ad61df8-aa1b-4430-888d-1428e935ccd7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390351, 'name': SuspendVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.127974] env[62965]: DEBUG nova.scheduler.client.report [None req-eac9b66d-1384-4e35-8f5d-c3800838412a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1056.307248] env[62965]: DEBUG nova.compute.manager [req-6cebf867-b4ff-4532-9e8a-f850a9bb3b67 req-18cac1e5-e832-47a4-a7e4-2eb700852414 service nova] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Received event network-changed-5f939a36-bc5f-4fd7-b29e-011f72b3256c {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1056.307653] env[62965]: DEBUG nova.compute.manager [req-6cebf867-b4ff-4532-9e8a-f850a9bb3b67 req-18cac1e5-e832-47a4-a7e4-2eb700852414 service nova] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Refreshing instance network info cache due to event network-changed-5f939a36-bc5f-4fd7-b29e-011f72b3256c. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 1056.307900] env[62965]: DEBUG oslo_concurrency.lockutils [req-6cebf867-b4ff-4532-9e8a-f850a9bb3b67 req-18cac1e5-e832-47a4-a7e4-2eb700852414 service nova] Acquiring lock "refresh_cache-73a7ca0e-b1d3-4464-969b-5111d2506a8b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1056.308066] env[62965]: DEBUG oslo_concurrency.lockutils [req-6cebf867-b4ff-4532-9e8a-f850a9bb3b67 req-18cac1e5-e832-47a4-a7e4-2eb700852414 service nova] Acquired lock "refresh_cache-73a7ca0e-b1d3-4464-969b-5111d2506a8b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.308241] env[62965]: DEBUG nova.network.neutron [req-6cebf867-b4ff-4532-9e8a-f850a9bb3b67 req-18cac1e5-e832-47a4-a7e4-2eb700852414 service nova] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Refreshing network info cache for port 5f939a36-bc5f-4fd7-b29e-011f72b3256c {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1056.588446] env[62965]: DEBUG oslo_vmware.api [None req-6ad61df8-aa1b-4430-888d-1428e935ccd7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390351, 'name': SuspendVM_Task} progress is 62%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.020250] env[62965]: DEBUG nova.network.neutron [req-6cebf867-b4ff-4532-9e8a-f850a9bb3b67 req-18cac1e5-e832-47a4-a7e4-2eb700852414 service nova] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Updated VIF entry in instance network info cache for port 5f939a36-bc5f-4fd7-b29e-011f72b3256c. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1057.020661] env[62965]: DEBUG nova.network.neutron [req-6cebf867-b4ff-4532-9e8a-f850a9bb3b67 req-18cac1e5-e832-47a4-a7e4-2eb700852414 service nova] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Updating instance_info_cache with network_info: [{"id": "5f939a36-bc5f-4fd7-b29e-011f72b3256c", "address": "fa:16:3e:e3:06:99", "network": {"id": "49dd42b0-6397-4fdd-a319-ed920c72a5af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-373172566-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d91c99b14018469e8998fd1ccc0fab4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f939a36-bc", "ovs_interfaceid": "5f939a36-bc5f-4fd7-b29e-011f72b3256c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.089135] env[62965]: DEBUG oslo_vmware.api [None req-6ad61df8-aa1b-4430-888d-1428e935ccd7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390351, 'name': SuspendVM_Task, 'duration_secs': 0.621814} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.089448] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-6ad61df8-aa1b-4430-888d-1428e935ccd7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Suspended the VM {{(pid=62965) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1057.089640] env[62965]: DEBUG nova.compute.manager [None req-6ad61df8-aa1b-4430-888d-1428e935ccd7 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1057.090473] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92f171c4-95e8-46e9-83ed-6c39e3997839 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.139234] env[62965]: DEBUG oslo_concurrency.lockutils [None req-eac9b66d-1384-4e35-8f5d-c3800838412a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.193s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.142150] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f91cb700-c247-47d4-a041-bf4e04bfb125 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.805s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.142398] env[62965]: DEBUG nova.objects.instance [None req-f91cb700-c247-47d4-a041-bf4e04bfb125 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lazy-loading 'resources' on Instance uuid 58c62ac1-203c-4370-8c94-420d655b7c08 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1057.497710] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1057.497958] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1057.498125] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Starting heal instance info cache {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10257}} [ 1057.523026] env[62965]: DEBUG oslo_concurrency.lockutils [req-6cebf867-b4ff-4532-9e8a-f850a9bb3b67 req-18cac1e5-e832-47a4-a7e4-2eb700852414 service nova] Releasing lock "refresh_cache-73a7ca0e-b1d3-4464-969b-5111d2506a8b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1057.699660] env[62965]: INFO nova.scheduler.client.report [None req-eac9b66d-1384-4e35-8f5d-c3800838412a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Deleted allocation for migration a1de2151-bfa9-4136-9a84-6811eb0921cc [ 1057.752420] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9560ff96-b43d-4f53-900b-258229f8d57d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.760333] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76700854-4682-46cb-afab-37cc6ed6a153 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.791026] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f69df0df-8d9e-49cb-b528-3a3d5cad7f37 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.798323] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c5cfb95-717d-478d-bb91-57548c254f1a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.811775] env[62965]: DEBUG nova.compute.provider_tree [None req-f91cb700-c247-47d4-a041-bf4e04bfb125 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1058.028081] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Acquiring lock "refresh_cache-bd36601b-5a70-4a29-8ef1-d1e925f41de7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.028081] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Acquired lock "refresh_cache-bd36601b-5a70-4a29-8ef1-d1e925f41de7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.028081] env[62965]: DEBUG nova.network.neutron [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Forcefully refreshing network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1058.205202] env[62965]: DEBUG oslo_concurrency.lockutils [None req-eac9b66d-1384-4e35-8f5d-c3800838412a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "c2448b76-1553-4b68-a731-0bb0967f4c1d" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.658s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.315199] env[62965]: DEBUG nova.scheduler.client.report [None req-f91cb700-c247-47d4-a041-bf4e04bfb125 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1058.449359] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8cb355f9-03a2-426b-8a8d-7f484e97e9a6 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "88e45c29-4d21-4f91-92f4-fdebaf905ccb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.449786] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8cb355f9-03a2-426b-8a8d-7f484e97e9a6 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "88e45c29-4d21-4f91-92f4-fdebaf905ccb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.450067] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8cb355f9-03a2-426b-8a8d-7f484e97e9a6 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "88e45c29-4d21-4f91-92f4-fdebaf905ccb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.450299] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8cb355f9-03a2-426b-8a8d-7f484e97e9a6 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "88e45c29-4d21-4f91-92f4-fdebaf905ccb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.450482] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8cb355f9-03a2-426b-8a8d-7f484e97e9a6 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "88e45c29-4d21-4f91-92f4-fdebaf905ccb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.452964] env[62965]: INFO nova.compute.manager [None req-8cb355f9-03a2-426b-8a8d-7f484e97e9a6 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Terminating instance [ 1058.736314] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8b274e64-4432-4afb-b2cd-e89b1c484725 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "c2448b76-1553-4b68-a731-0bb0967f4c1d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.736675] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8b274e64-4432-4afb-b2cd-e89b1c484725 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "c2448b76-1553-4b68-a731-0bb0967f4c1d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.736722] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8b274e64-4432-4afb-b2cd-e89b1c484725 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "c2448b76-1553-4b68-a731-0bb0967f4c1d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.736905] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8b274e64-4432-4afb-b2cd-e89b1c484725 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "c2448b76-1553-4b68-a731-0bb0967f4c1d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.737089] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8b274e64-4432-4afb-b2cd-e89b1c484725 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "c2448b76-1553-4b68-a731-0bb0967f4c1d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.739232] env[62965]: INFO nova.compute.manager [None req-8b274e64-4432-4afb-b2cd-e89b1c484725 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Terminating instance [ 1058.820206] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f91cb700-c247-47d4-a041-bf4e04bfb125 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.678s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.822427] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.283s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.823885] env[62965]: INFO nova.compute.claims [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1058.837848] env[62965]: INFO nova.scheduler.client.report [None req-f91cb700-c247-47d4-a041-bf4e04bfb125 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Deleted allocations for instance 58c62ac1-203c-4370-8c94-420d655b7c08 [ 1058.957084] env[62965]: DEBUG nova.compute.manager [None req-8cb355f9-03a2-426b-8a8d-7f484e97e9a6 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1058.957333] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-8cb355f9-03a2-426b-8a8d-7f484e97e9a6 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1058.958241] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4052547c-d0b0-45dc-9034-811a63bfec62 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.966230] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-8cb355f9-03a2-426b-8a8d-7f484e97e9a6 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1058.966713] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e2a268be-0ec3-4de0-a515-b3c9da6670b9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.042126] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-8cb355f9-03a2-426b-8a8d-7f484e97e9a6 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1059.042378] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-8cb355f9-03a2-426b-8a8d-7f484e97e9a6 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Deleting contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1059.042574] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-8cb355f9-03a2-426b-8a8d-7f484e97e9a6 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Deleting the datastore file [datastore2] 88e45c29-4d21-4f91-92f4-fdebaf905ccb {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1059.042841] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0c15b7cd-6793-4907-9f34-1f0bc1861468 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.049332] env[62965]: DEBUG oslo_vmware.api [None req-8cb355f9-03a2-426b-8a8d-7f484e97e9a6 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 1059.049332] env[62965]: value = "task-1390353" [ 1059.049332] env[62965]: _type = "Task" [ 1059.049332] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.056889] env[62965]: DEBUG oslo_vmware.api [None req-8cb355f9-03a2-426b-8a8d-7f484e97e9a6 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390353, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.239948] env[62965]: DEBUG nova.network.neutron [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Updating instance_info_cache with network_info: [{"id": "7faa0a61-99a2-4ae4-b60a-ef2eba7aef31", "address": "fa:16:3e:ec:07:ba", "network": {"id": "a0a5e605-d580-4b7b-b3c0-5c7395bd5a8b", "bridge": "br-int", "label": "tempest-ServersTestJSON-1679618017-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1ed71548c085499981fa1b102d660368", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0cd5d325-3053-407e-a4ee-f627e82a23f9", "external-id": "nsx-vlan-transportzone-809", "segmentation_id": 809, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7faa0a61-99", "ovs_interfaceid": "7faa0a61-99a2-4ae4-b60a-ef2eba7aef31", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1059.243135] env[62965]: DEBUG nova.compute.manager [None req-8b274e64-4432-4afb-b2cd-e89b1c484725 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1059.243347] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-8b274e64-4432-4afb-b2cd-e89b1c484725 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1059.244214] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12f7bfbb-c34f-4bd4-8dec-3cc57e4f4431 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.252170] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b274e64-4432-4afb-b2cd-e89b1c484725 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1059.252421] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9a4419ba-8350-4fe9-aed7-47f29f589099 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.259517] env[62965]: DEBUG oslo_vmware.api [None req-8b274e64-4432-4afb-b2cd-e89b1c484725 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1059.259517] env[62965]: value = "task-1390354" [ 1059.259517] env[62965]: _type = "Task" [ 1059.259517] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.267704] env[62965]: DEBUG oslo_vmware.api [None req-8b274e64-4432-4afb-b2cd-e89b1c484725 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390354, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.344378] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f91cb700-c247-47d4-a041-bf4e04bfb125 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "58c62ac1-203c-4370-8c94-420d655b7c08" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.011s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.558813] env[62965]: DEBUG oslo_vmware.api [None req-8cb355f9-03a2-426b-8a8d-7f484e97e9a6 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390353, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141126} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.559098] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-8cb355f9-03a2-426b-8a8d-7f484e97e9a6 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1059.559301] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-8cb355f9-03a2-426b-8a8d-7f484e97e9a6 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Deleted contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1059.559467] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-8cb355f9-03a2-426b-8a8d-7f484e97e9a6 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1059.559649] env[62965]: INFO nova.compute.manager [None req-8cb355f9-03a2-426b-8a8d-7f484e97e9a6 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Took 0.60 seconds to destroy the instance on the hypervisor. [ 1059.559897] env[62965]: DEBUG oslo.service.loopingcall [None req-8cb355f9-03a2-426b-8a8d-7f484e97e9a6 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1059.560115] env[62965]: DEBUG nova.compute.manager [-] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1059.560210] env[62965]: DEBUG nova.network.neutron [-] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1059.743096] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Releasing lock "refresh_cache-bd36601b-5a70-4a29-8ef1-d1e925f41de7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.743430] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Updated the network info_cache for instance {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10328}} [ 1059.743574] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1059.743737] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1059.743923] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1059.744093] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1059.744240] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1059.744386] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1059.744513] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62965) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10876}} [ 1059.744659] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1059.770135] env[62965]: DEBUG oslo_vmware.api [None req-8b274e64-4432-4afb-b2cd-e89b1c484725 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390354, 'name': PowerOffVM_Task, 'duration_secs': 0.154372} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.770410] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b274e64-4432-4afb-b2cd-e89b1c484725 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1059.770606] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-8b274e64-4432-4afb-b2cd-e89b1c484725 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1059.771365] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d91ae71d-a91a-4589-b483-ad621754a14b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.812958] env[62965]: DEBUG nova.compute.manager [req-526b607a-a693-4de9-9640-b3f1de26bdf3 req-89581779-f206-4175-ac65-6907eff2f52f service nova] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Received event network-vif-deleted-52432dbd-16ac-41f2-8bee-6d27c80125dd {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1059.813182] env[62965]: INFO nova.compute.manager [req-526b607a-a693-4de9-9640-b3f1de26bdf3 req-89581779-f206-4175-ac65-6907eff2f52f service nova] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Neutron deleted interface 52432dbd-16ac-41f2-8bee-6d27c80125dd; detaching it from the instance and deleting it from the info cache [ 1059.813438] env[62965]: DEBUG nova.network.neutron [req-526b607a-a693-4de9-9640-b3f1de26bdf3 req-89581779-f206-4175-ac65-6907eff2f52f service nova] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1059.830726] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-8b274e64-4432-4afb-b2cd-e89b1c484725 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1059.830946] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-8b274e64-4432-4afb-b2cd-e89b1c484725 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Deleting contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1059.831156] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b274e64-4432-4afb-b2cd-e89b1c484725 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Deleting the datastore file [datastore1] c2448b76-1553-4b68-a731-0bb0967f4c1d {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1059.834903] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-35d6bed0-4f16-429c-b1eb-e8123b90e5c9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.841770] env[62965]: DEBUG oslo_vmware.api [None req-8b274e64-4432-4afb-b2cd-e89b1c484725 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1059.841770] env[62965]: value = "task-1390356" [ 1059.841770] env[62965]: _type = "Task" [ 1059.841770] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.849976] env[62965]: DEBUG oslo_vmware.api [None req-8b274e64-4432-4afb-b2cd-e89b1c484725 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390356, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.916926] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38a771bd-da89-49a8-8106-9bc5ef5f9cae {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.924533] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e50cf5ed-6b4a-4eee-a26c-f4c47b534f59 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.955089] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db4bc87f-afdd-4e90-876e-3fbcfa602739 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.962894] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26bdf8f2-8b38-4167-95bc-80f12a976d34 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.976376] env[62965]: DEBUG nova.compute.provider_tree [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1060.247884] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.253539] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b5933398-25ce-464a-9644-c5a41b2887d8 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "bd36601b-5a70-4a29-8ef1-d1e925f41de7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.253779] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b5933398-25ce-464a-9644-c5a41b2887d8 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "bd36601b-5a70-4a29-8ef1-d1e925f41de7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.253991] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b5933398-25ce-464a-9644-c5a41b2887d8 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "bd36601b-5a70-4a29-8ef1-d1e925f41de7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.254199] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b5933398-25ce-464a-9644-c5a41b2887d8 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "bd36601b-5a70-4a29-8ef1-d1e925f41de7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.254379] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b5933398-25ce-464a-9644-c5a41b2887d8 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "bd36601b-5a70-4a29-8ef1-d1e925f41de7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.256621] env[62965]: INFO nova.compute.manager [None req-b5933398-25ce-464a-9644-c5a41b2887d8 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Terminating instance [ 1060.291717] env[62965]: DEBUG nova.network.neutron [-] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.316572] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-56b65bdd-60c0-41f2-9b4b-bd5fc2a13e98 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.326438] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01993035-48fa-4dd9-b80e-bb2c6cb4e649 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.351778] env[62965]: DEBUG nova.compute.manager [req-526b607a-a693-4de9-9640-b3f1de26bdf3 req-89581779-f206-4175-ac65-6907eff2f52f service nova] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Detach interface failed, port_id=52432dbd-16ac-41f2-8bee-6d27c80125dd, reason: Instance 88e45c29-4d21-4f91-92f4-fdebaf905ccb could not be found. {{(pid=62965) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11294}} [ 1060.357271] env[62965]: DEBUG oslo_vmware.api [None req-8b274e64-4432-4afb-b2cd-e89b1c484725 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390356, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146453} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.357593] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b274e64-4432-4afb-b2cd-e89b1c484725 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1060.357806] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-8b274e64-4432-4afb-b2cd-e89b1c484725 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Deleted contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1060.358163] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-8b274e64-4432-4afb-b2cd-e89b1c484725 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1060.358234] env[62965]: INFO nova.compute.manager [None req-8b274e64-4432-4afb-b2cd-e89b1c484725 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1060.358508] env[62965]: DEBUG oslo.service.loopingcall [None req-8b274e64-4432-4afb-b2cd-e89b1c484725 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1060.358858] env[62965]: DEBUG nova.compute.manager [-] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1060.358858] env[62965]: DEBUG nova.network.neutron [-] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1060.479401] env[62965]: DEBUG nova.scheduler.client.report [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1060.760369] env[62965]: DEBUG nova.compute.manager [None req-b5933398-25ce-464a-9644-c5a41b2887d8 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1060.760693] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b5933398-25ce-464a-9644-c5a41b2887d8 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1060.761563] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9980351-2bf1-4499-b4ba-2df6eeed8ced {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.768823] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5933398-25ce-464a-9644-c5a41b2887d8 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1060.769076] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-41f24699-5586-4c2f-af03-f3ce8a1d9298 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.774445] env[62965]: DEBUG oslo_vmware.api [None req-b5933398-25ce-464a-9644-c5a41b2887d8 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 1060.774445] env[62965]: value = "task-1390357" [ 1060.774445] env[62965]: _type = "Task" [ 1060.774445] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.782572] env[62965]: DEBUG oslo_vmware.api [None req-b5933398-25ce-464a-9644-c5a41b2887d8 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390357, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.794229] env[62965]: INFO nova.compute.manager [-] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Took 1.23 seconds to deallocate network for instance. [ 1060.986019] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.162s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.986865] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.739s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.987052] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.987207] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62965) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1060.988076] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73cbb318-c4e4-4aec-a76a-d487cbda59ea {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.996410] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db243078-c75b-426a-a03f-3b2d60521282 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.010820] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e63124a4-82d8-4b9d-9a09-d6dca9709302 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.017167] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b2d25be-9961-459f-b09b-889270a5227b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.049896] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180228MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62965) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1061.050068] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.050273] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.284855] env[62965]: DEBUG oslo_vmware.api [None req-b5933398-25ce-464a-9644-c5a41b2887d8 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390357, 'name': PowerOffVM_Task, 'duration_secs': 0.213554} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.285174] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5933398-25ce-464a-9644-c5a41b2887d8 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1061.285354] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b5933398-25ce-464a-9644-c5a41b2887d8 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1061.285624] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d1e38a3a-b931-4790-9ea7-0830c5fc1925 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.301480] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8cb355f9-03a2-426b-8a8d-7f484e97e9a6 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.301828] env[62965]: DEBUG nova.network.neutron [-] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.386400] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b5933398-25ce-464a-9644-c5a41b2887d8 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1061.386639] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b5933398-25ce-464a-9644-c5a41b2887d8 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Deleting contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1061.386828] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5933398-25ce-464a-9644-c5a41b2887d8 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Deleting the datastore file [datastore1] bd36601b-5a70-4a29-8ef1-d1e925f41de7 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1061.387114] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b672f602-93f0-4407-9740-d75f1a5bc78d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.394526] env[62965]: DEBUG oslo_vmware.api [None req-b5933398-25ce-464a-9644-c5a41b2887d8 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for the task: (returnval){ [ 1061.394526] env[62965]: value = "task-1390359" [ 1061.394526] env[62965]: _type = "Task" [ 1061.394526] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.403484] env[62965]: DEBUG oslo_vmware.api [None req-b5933398-25ce-464a-9644-c5a41b2887d8 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390359, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.488239] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Acquiring lock "fb13801c-3bb2-4be2-9ea1-8feeff858d98" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.488530] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Lock "fb13801c-3bb2-4be2-9ea1-8feeff858d98" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.804744] env[62965]: INFO nova.compute.manager [-] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Took 1.45 seconds to deallocate network for instance. [ 1061.841372] env[62965]: DEBUG nova.compute.manager [req-3763c7d0-c9f9-446f-9fc7-5e8014057963 req-e046291f-f140-4cb8-8080-0ab3b64a0df4 service nova] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Received event network-vif-deleted-57672e8b-1df1-463d-b9b3-8290667d9fd2 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1061.906352] env[62965]: DEBUG oslo_vmware.api [None req-b5933398-25ce-464a-9644-c5a41b2887d8 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Task: {'id': task-1390359, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.119544} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.906659] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5933398-25ce-464a-9644-c5a41b2887d8 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1061.906914] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b5933398-25ce-464a-9644-c5a41b2887d8 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Deleted contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1061.907160] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b5933398-25ce-464a-9644-c5a41b2887d8 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1061.907377] env[62965]: INFO nova.compute.manager [None req-b5933398-25ce-464a-9644-c5a41b2887d8 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1061.907664] env[62965]: DEBUG oslo.service.loopingcall [None req-b5933398-25ce-464a-9644-c5a41b2887d8 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1061.908169] env[62965]: DEBUG nova.compute.manager [-] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1061.908303] env[62965]: DEBUG nova.network.neutron [-] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1061.991638] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Lock "fb13801c-3bb2-4be2-9ea1-8feeff858d98" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.503s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.992186] env[62965]: DEBUG nova.compute.manager [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1062.077143] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 924c5cdc-0450-4ced-b920-50e2d5060fd2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1062.078130] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance bd36601b-5a70-4a29-8ef1-d1e925f41de7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1062.078130] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance c2448b76-1553-4b68-a731-0bb0967f4c1d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1062.078130] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 73a7ca0e-b1d3-4464-969b-5111d2506a8b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1062.078130] env[62965]: WARNING nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 88e45c29-4d21-4f91-92f4-fdebaf905ccb is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1062.078309] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 740b4d83-afc5-4ee2-9130-8c1d0697e5e5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1062.078525] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=62965) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1062.078746] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1536MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=62965) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1062.171696] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2a162de-f96f-4f6b-9f35-77c18a798694 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.179259] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5791c957-ef3b-4b28-a12c-b12fe9807679 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.210380] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d399aae-b756-44a5-bb3a-a188f9ec27cd {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.217635] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eb639ca-2e4a-4cbc-9d0a-86acfdebbc90 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.230677] env[62965]: DEBUG nova.compute.provider_tree [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1062.310453] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8b274e64-4432-4afb-b2cd-e89b1c484725 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.497469] env[62965]: DEBUG nova.compute.utils [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1062.499239] env[62965]: DEBUG nova.compute.manager [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1062.499473] env[62965]: DEBUG nova.network.neutron [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1062.545329] env[62965]: DEBUG nova.policy [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ad38aeec07df4a5e811609db39f9fbfe', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a3b9b66397564d288f333fb20686894e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 1062.655841] env[62965]: DEBUG nova.network.neutron [-] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.733868] env[62965]: DEBUG nova.scheduler.client.report [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1062.789223] env[62965]: DEBUG nova.network.neutron [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Successfully created port: 19aed2fc-7ec6-4240-9c8e-a109f243d2f5 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1063.002352] env[62965]: DEBUG nova.compute.manager [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1063.158594] env[62965]: INFO nova.compute.manager [-] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Took 1.25 seconds to deallocate network for instance. [ 1063.238871] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62965) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1063.239083] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.189s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.239379] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8cb355f9-03a2-426b-8a8d-7f484e97e9a6 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.938s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.239574] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8cb355f9-03a2-426b-8a8d-7f484e97e9a6 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.241613] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8b274e64-4432-4afb-b2cd-e89b1c484725 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.931s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.241837] env[62965]: DEBUG nova.objects.instance [None req-8b274e64-4432-4afb-b2cd-e89b1c484725 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lazy-loading 'resources' on Instance uuid c2448b76-1553-4b68-a731-0bb0967f4c1d {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1063.262687] env[62965]: INFO nova.scheduler.client.report [None req-8cb355f9-03a2-426b-8a8d-7f484e97e9a6 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Deleted allocations for instance 88e45c29-4d21-4f91-92f4-fdebaf905ccb [ 1063.664886] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b5933398-25ce-464a-9644-c5a41b2887d8 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.769487] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8cb355f9-03a2-426b-8a8d-7f484e97e9a6 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "88e45c29-4d21-4f91-92f4-fdebaf905ccb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.320s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.815511] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b3096be-969b-488e-89ce-c2b0d93b19ad {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.823668] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dce56aa-79e5-4486-a66f-f50501ea136e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.852956] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53562dfa-c1f0-4083-9785-00ef1eeaabed {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.861027] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17536606-36ee-4fbf-8350-8cda7d1d827e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.867116] env[62965]: DEBUG nova.compute.manager [req-0ccf69b2-c82f-4755-96d4-63979ad34a8f req-7fa8b957-6e75-489b-9ce1-b6c3b4eed80e service nova] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Received event network-vif-deleted-7faa0a61-99a2-4ae4-b60a-ef2eba7aef31 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1063.875225] env[62965]: DEBUG nova.compute.provider_tree [None req-8b274e64-4432-4afb-b2cd-e89b1c484725 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1064.011631] env[62965]: DEBUG nova.compute.manager [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1064.032309] env[62965]: DEBUG nova.virt.hardware [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1064.032572] env[62965]: DEBUG nova.virt.hardware [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1064.032734] env[62965]: DEBUG nova.virt.hardware [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1064.032982] env[62965]: DEBUG nova.virt.hardware [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1064.033078] env[62965]: DEBUG nova.virt.hardware [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1064.033232] env[62965]: DEBUG nova.virt.hardware [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1064.033440] env[62965]: DEBUG nova.virt.hardware [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1064.033605] env[62965]: DEBUG nova.virt.hardware [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1064.033779] env[62965]: DEBUG nova.virt.hardware [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1064.033942] env[62965]: DEBUG nova.virt.hardware [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1064.034136] env[62965]: DEBUG nova.virt.hardware [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1064.034988] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12a20551-101f-45cf-a69e-cb6773d1eb0d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.042947] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92cbaa12-0347-4187-862d-5b5fea77de50 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.237401] env[62965]: DEBUG nova.network.neutron [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Successfully updated port: 19aed2fc-7ec6-4240-9c8e-a109f243d2f5 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1064.378495] env[62965]: DEBUG nova.scheduler.client.report [None req-8b274e64-4432-4afb-b2cd-e89b1c484725 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1064.598360] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.598681] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.739892] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Acquiring lock "refresh_cache-740b4d83-afc5-4ee2-9130-8c1d0697e5e5" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.740055] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Acquired lock "refresh_cache-740b4d83-afc5-4ee2-9130-8c1d0697e5e5" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.740208] env[62965]: DEBUG nova.network.neutron [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1064.883624] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8b274e64-4432-4afb-b2cd-e89b1c484725 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.642s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.885970] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b5933398-25ce-464a-9644-c5a41b2887d8 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.221s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.886221] env[62965]: DEBUG nova.objects.instance [None req-b5933398-25ce-464a-9644-c5a41b2887d8 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lazy-loading 'resources' on Instance uuid bd36601b-5a70-4a29-8ef1-d1e925f41de7 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1064.903582] env[62965]: INFO nova.scheduler.client.report [None req-8b274e64-4432-4afb-b2cd-e89b1c484725 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Deleted allocations for instance c2448b76-1553-4b68-a731-0bb0967f4c1d [ 1065.100848] env[62965]: DEBUG nova.compute.manager [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1065.272988] env[62965]: DEBUG nova.network.neutron [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1065.404442] env[62965]: DEBUG nova.network.neutron [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Updating instance_info_cache with network_info: [{"id": "19aed2fc-7ec6-4240-9c8e-a109f243d2f5", "address": "fa:16:3e:30:0a:ba", "network": {"id": "63ba984a-91a8-45ba-8daa-817e8eddc223", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1459561773-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a3b9b66397564d288f333fb20686894e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19aed2fc-7e", "ovs_interfaceid": "19aed2fc-7ec6-4240-9c8e-a109f243d2f5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1065.410648] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8b274e64-4432-4afb-b2cd-e89b1c484725 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "c2448b76-1553-4b68-a731-0bb0967f4c1d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.674s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.456389] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bd62943-8941-44fc-b73d-16488bb80215 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.465191] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-789714f1-fa09-4842-b7b3-5838ec093937 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.496971] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b470abf-f592-4e3d-b89a-15725557e6ea {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.504042] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f81ffdb-b2cd-40d3-ac4e-a049c0e76f1b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.516669] env[62965]: DEBUG nova.compute.provider_tree [None req-b5933398-25ce-464a-9644-c5a41b2887d8 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1065.620948] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.891257] env[62965]: DEBUG nova.compute.manager [req-4a1bf7dd-006c-4ce7-9eea-307943b1e385 req-46a8d481-a9d1-4e5b-90c1-9978153bf52e service nova] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Received event network-vif-plugged-19aed2fc-7ec6-4240-9c8e-a109f243d2f5 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1065.891509] env[62965]: DEBUG oslo_concurrency.lockutils [req-4a1bf7dd-006c-4ce7-9eea-307943b1e385 req-46a8d481-a9d1-4e5b-90c1-9978153bf52e service nova] Acquiring lock "740b4d83-afc5-4ee2-9130-8c1d0697e5e5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.891737] env[62965]: DEBUG oslo_concurrency.lockutils [req-4a1bf7dd-006c-4ce7-9eea-307943b1e385 req-46a8d481-a9d1-4e5b-90c1-9978153bf52e service nova] Lock "740b4d83-afc5-4ee2-9130-8c1d0697e5e5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.891913] env[62965]: DEBUG oslo_concurrency.lockutils [req-4a1bf7dd-006c-4ce7-9eea-307943b1e385 req-46a8d481-a9d1-4e5b-90c1-9978153bf52e service nova] Lock "740b4d83-afc5-4ee2-9130-8c1d0697e5e5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.892096] env[62965]: DEBUG nova.compute.manager [req-4a1bf7dd-006c-4ce7-9eea-307943b1e385 req-46a8d481-a9d1-4e5b-90c1-9978153bf52e service nova] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] No waiting events found dispatching network-vif-plugged-19aed2fc-7ec6-4240-9c8e-a109f243d2f5 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1065.892272] env[62965]: WARNING nova.compute.manager [req-4a1bf7dd-006c-4ce7-9eea-307943b1e385 req-46a8d481-a9d1-4e5b-90c1-9978153bf52e service nova] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Received unexpected event network-vif-plugged-19aed2fc-7ec6-4240-9c8e-a109f243d2f5 for instance with vm_state building and task_state spawning. [ 1065.892437] env[62965]: DEBUG nova.compute.manager [req-4a1bf7dd-006c-4ce7-9eea-307943b1e385 req-46a8d481-a9d1-4e5b-90c1-9978153bf52e service nova] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Received event network-changed-19aed2fc-7ec6-4240-9c8e-a109f243d2f5 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1065.892604] env[62965]: DEBUG nova.compute.manager [req-4a1bf7dd-006c-4ce7-9eea-307943b1e385 req-46a8d481-a9d1-4e5b-90c1-9978153bf52e service nova] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Refreshing instance network info cache due to event network-changed-19aed2fc-7ec6-4240-9c8e-a109f243d2f5. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 1065.892801] env[62965]: DEBUG oslo_concurrency.lockutils [req-4a1bf7dd-006c-4ce7-9eea-307943b1e385 req-46a8d481-a9d1-4e5b-90c1-9978153bf52e service nova] Acquiring lock "refresh_cache-740b4d83-afc5-4ee2-9130-8c1d0697e5e5" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.907193] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Releasing lock "refresh_cache-740b4d83-afc5-4ee2-9130-8c1d0697e5e5" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.907483] env[62965]: DEBUG nova.compute.manager [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Instance network_info: |[{"id": "19aed2fc-7ec6-4240-9c8e-a109f243d2f5", "address": "fa:16:3e:30:0a:ba", "network": {"id": "63ba984a-91a8-45ba-8daa-817e8eddc223", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1459561773-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a3b9b66397564d288f333fb20686894e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19aed2fc-7e", "ovs_interfaceid": "19aed2fc-7ec6-4240-9c8e-a109f243d2f5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1065.907769] env[62965]: DEBUG oslo_concurrency.lockutils [req-4a1bf7dd-006c-4ce7-9eea-307943b1e385 req-46a8d481-a9d1-4e5b-90c1-9978153bf52e service nova] Acquired lock "refresh_cache-740b4d83-afc5-4ee2-9130-8c1d0697e5e5" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.907973] env[62965]: DEBUG nova.network.neutron [req-4a1bf7dd-006c-4ce7-9eea-307943b1e385 req-46a8d481-a9d1-4e5b-90c1-9978153bf52e service nova] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Refreshing network info cache for port 19aed2fc-7ec6-4240-9c8e-a109f243d2f5 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1065.909400] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:30:0a:ba', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '19aed2fc-7ec6-4240-9c8e-a109f243d2f5', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1065.916811] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Creating folder: Project (a3b9b66397564d288f333fb20686894e). Parent ref: group-v294931. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1065.920168] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-82014d57-f910-4a8e-840d-d2754592d347 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.934261] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Created folder: Project (a3b9b66397564d288f333fb20686894e) in parent group-v294931. [ 1065.934261] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Creating folder: Instances. Parent ref: group-v295055. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1065.934464] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7a332fa8-a425-405a-b7e6-9782d61c17b8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.944802] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Created folder: Instances in parent group-v295055. [ 1065.944897] env[62965]: DEBUG oslo.service.loopingcall [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1065.945324] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1065.945790] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-15f6ac4a-8abe-45c5-bbb6-ce784614af29 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.968372] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1065.968372] env[62965]: value = "task-1390362" [ 1065.968372] env[62965]: _type = "Task" [ 1065.968372] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.975984] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390362, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.020097] env[62965]: DEBUG nova.scheduler.client.report [None req-b5933398-25ce-464a-9644-c5a41b2887d8 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1066.144489] env[62965]: DEBUG nova.network.neutron [req-4a1bf7dd-006c-4ce7-9eea-307943b1e385 req-46a8d481-a9d1-4e5b-90c1-9978153bf52e service nova] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Updated VIF entry in instance network info cache for port 19aed2fc-7ec6-4240-9c8e-a109f243d2f5. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1066.144868] env[62965]: DEBUG nova.network.neutron [req-4a1bf7dd-006c-4ce7-9eea-307943b1e385 req-46a8d481-a9d1-4e5b-90c1-9978153bf52e service nova] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Updating instance_info_cache with network_info: [{"id": "19aed2fc-7ec6-4240-9c8e-a109f243d2f5", "address": "fa:16:3e:30:0a:ba", "network": {"id": "63ba984a-91a8-45ba-8daa-817e8eddc223", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1459561773-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a3b9b66397564d288f333fb20686894e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19aed2fc-7e", "ovs_interfaceid": "19aed2fc-7ec6-4240-9c8e-a109f243d2f5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.478608] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390362, 'name': CreateVM_Task, 'duration_secs': 0.319258} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.478608] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1066.479014] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1066.479201] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.479589] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1066.479858] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7901d1c6-0ae1-49dd-a9d3-807a2f8f360a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.484097] env[62965]: DEBUG oslo_vmware.api [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Waiting for the task: (returnval){ [ 1066.484097] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]529e36eb-4b9c-de1d-4a59-cdac04fd4849" [ 1066.484097] env[62965]: _type = "Task" [ 1066.484097] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.491078] env[62965]: DEBUG oslo_vmware.api [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]529e36eb-4b9c-de1d-4a59-cdac04fd4849, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.526986] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b5933398-25ce-464a-9644-c5a41b2887d8 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.641s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.529061] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.908s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.530553] env[62965]: INFO nova.compute.claims [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1066.544828] env[62965]: INFO nova.scheduler.client.report [None req-b5933398-25ce-464a-9644-c5a41b2887d8 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Deleted allocations for instance bd36601b-5a70-4a29-8ef1-d1e925f41de7 [ 1066.648034] env[62965]: DEBUG oslo_concurrency.lockutils [req-4a1bf7dd-006c-4ce7-9eea-307943b1e385 req-46a8d481-a9d1-4e5b-90c1-9978153bf52e service nova] Releasing lock "refresh_cache-740b4d83-afc5-4ee2-9130-8c1d0697e5e5" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.824553] env[62965]: DEBUG oslo_concurrency.lockutils [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "385648d7-20e3-442c-b6a1-3d972d266487" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.824799] env[62965]: DEBUG oslo_concurrency.lockutils [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "385648d7-20e3-442c-b6a1-3d972d266487" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.993784] env[62965]: DEBUG oslo_vmware.api [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]529e36eb-4b9c-de1d-4a59-cdac04fd4849, 'name': SearchDatastore_Task, 'duration_secs': 0.014386} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.994052] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.994283] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1066.994516] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1066.994667] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.994843] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1066.995153] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e07b588f-ed06-44c2-b685-241160fd7eda {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.002714] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1067.002893] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1067.003576] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ef4911c-ab69-420b-98b1-7b4faa329b59 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.008155] env[62965]: DEBUG oslo_vmware.api [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Waiting for the task: (returnval){ [ 1067.008155] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52a387c1-40ee-eb0e-9230-18626957bd5b" [ 1067.008155] env[62965]: _type = "Task" [ 1067.008155] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.015511] env[62965]: DEBUG oslo_vmware.api [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52a387c1-40ee-eb0e-9230-18626957bd5b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.050913] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b5933398-25ce-464a-9644-c5a41b2887d8 tempest-ServersTestJSON-756955642 tempest-ServersTestJSON-756955642-project-member] Lock "bd36601b-5a70-4a29-8ef1-d1e925f41de7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.797s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.327307] env[62965]: DEBUG nova.compute.manager [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1067.519136] env[62965]: DEBUG oslo_vmware.api [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52a387c1-40ee-eb0e-9230-18626957bd5b, 'name': SearchDatastore_Task, 'duration_secs': 0.007863} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.519946] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d863fe1-dba0-44ab-abc7-7120f2f7b9ad {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.524660] env[62965]: DEBUG oslo_vmware.api [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Waiting for the task: (returnval){ [ 1067.524660] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5250d472-1564-e7a0-d743-bf1e4d47f4c1" [ 1067.524660] env[62965]: _type = "Task" [ 1067.524660] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.531858] env[62965]: DEBUG oslo_vmware.api [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5250d472-1564-e7a0-d743-bf1e4d47f4c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.603150] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7cd57f5-8d40-4ef5-a904-6f17334ffef3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.610323] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d3bc90e-413f-4771-aafa-4b9b889b02db {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.641079] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8faeefdd-868a-44f8-b3d0-868fd49e5b7a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.648129] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-123ac56e-7fd4-4341-be19-ba339d4487d9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.660939] env[62965]: DEBUG nova.compute.provider_tree [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1067.847873] env[62965]: DEBUG oslo_concurrency.lockutils [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.034851] env[62965]: DEBUG oslo_vmware.api [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5250d472-1564-e7a0-d743-bf1e4d47f4c1, 'name': SearchDatastore_Task, 'duration_secs': 0.035534} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.035223] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.035510] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] 740b4d83-afc5-4ee2-9130-8c1d0697e5e5/740b4d83-afc5-4ee2-9130-8c1d0697e5e5.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1068.035784] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9aa51b6d-c426-4b0e-b671-3fe83c5d30a9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.042378] env[62965]: DEBUG oslo_vmware.api [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Waiting for the task: (returnval){ [ 1068.042378] env[62965]: value = "task-1390363" [ 1068.042378] env[62965]: _type = "Task" [ 1068.042378] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.049743] env[62965]: DEBUG oslo_vmware.api [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Task: {'id': task-1390363, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.168269] env[62965]: DEBUG nova.scheduler.client.report [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1068.554729] env[62965]: DEBUG oslo_vmware.api [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Task: {'id': task-1390363, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.673891] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.145s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.674502] env[62965]: DEBUG nova.compute.manager [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1068.677381] env[62965]: DEBUG oslo_concurrency.lockutils [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.830s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.678813] env[62965]: INFO nova.compute.claims [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1069.053150] env[62965]: DEBUG oslo_vmware.api [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Task: {'id': task-1390363, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.540177} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.053205] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] 740b4d83-afc5-4ee2-9130-8c1d0697e5e5/740b4d83-afc5-4ee2-9130-8c1d0697e5e5.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1069.053424] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1069.053683] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-400bccad-c8cc-4112-8ebc-71a91661d3ec {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.059981] env[62965]: DEBUG oslo_vmware.api [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Waiting for the task: (returnval){ [ 1069.059981] env[62965]: value = "task-1390364" [ 1069.059981] env[62965]: _type = "Task" [ 1069.059981] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.067483] env[62965]: DEBUG oslo_vmware.api [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Task: {'id': task-1390364, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.183636] env[62965]: DEBUG nova.compute.utils [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1069.184710] env[62965]: DEBUG nova.compute.manager [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1069.184921] env[62965]: DEBUG nova.network.neutron [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1069.225063] env[62965]: DEBUG nova.policy [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03c0a9a0ec3c46bf883c68d86cb816d6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '978666785b0743bb843de96585f96fb4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 1069.473137] env[62965]: DEBUG nova.network.neutron [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Successfully created port: 5232b21c-f849-45de-916a-788114415557 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1069.571366] env[62965]: DEBUG oslo_vmware.api [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Task: {'id': task-1390364, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.262689} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.571714] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1069.572648] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41a746ee-998c-4ab0-9d84-0c3b0eebac0e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.593851] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] 740b4d83-afc5-4ee2-9130-8c1d0697e5e5/740b4d83-afc5-4ee2-9130-8c1d0697e5e5.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1069.594166] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f10dc262-2308-4f10-8bff-2604fc579d6c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.615901] env[62965]: DEBUG oslo_vmware.api [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Waiting for the task: (returnval){ [ 1069.615901] env[62965]: value = "task-1390365" [ 1069.615901] env[62965]: _type = "Task" [ 1069.615901] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.623735] env[62965]: DEBUG oslo_vmware.api [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Task: {'id': task-1390365, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.689620] env[62965]: DEBUG nova.compute.manager [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1069.780921] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcc1c797-02f2-4c59-bbcd-a589936ee535 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.788381] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae0ff20e-a723-49da-afc8-3e2baa0f2654 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.817520] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5cb2aa2-476b-4ccd-ba1a-7f0840e56c6b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.824458] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a5454a5-eb84-4cd0-8099-c6cd7e62a75a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.837309] env[62965]: DEBUG nova.compute.provider_tree [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1070.068244] env[62965]: DEBUG oslo_concurrency.lockutils [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Acquiring lock "182e796f-2fbf-4304-9339-dc6e9b6b5839" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.068481] env[62965]: DEBUG oslo_concurrency.lockutils [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Lock "182e796f-2fbf-4304-9339-dc6e9b6b5839" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.125616] env[62965]: DEBUG oslo_vmware.api [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Task: {'id': task-1390365, 'name': ReconfigVM_Task, 'duration_secs': 0.268994} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.125911] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Reconfigured VM instance instance-00000068 to attach disk [datastore1] 740b4d83-afc5-4ee2-9130-8c1d0697e5e5/740b4d83-afc5-4ee2-9130-8c1d0697e5e5.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1070.126572] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3e46360d-f4e9-4b7b-9bfc-c6a5584e2777 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.133241] env[62965]: DEBUG oslo_vmware.api [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Waiting for the task: (returnval){ [ 1070.133241] env[62965]: value = "task-1390366" [ 1070.133241] env[62965]: _type = "Task" [ 1070.133241] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.141593] env[62965]: DEBUG oslo_vmware.api [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Task: {'id': task-1390366, 'name': Rename_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.340719] env[62965]: DEBUG nova.scheduler.client.report [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1070.570680] env[62965]: DEBUG nova.compute.manager [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1070.643380] env[62965]: DEBUG oslo_vmware.api [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Task: {'id': task-1390366, 'name': Rename_Task, 'duration_secs': 0.136238} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.643725] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1070.643832] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6f27cd4e-86ad-459b-9343-b68219caac04 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.649993] env[62965]: DEBUG oslo_vmware.api [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Waiting for the task: (returnval){ [ 1070.649993] env[62965]: value = "task-1390367" [ 1070.649993] env[62965]: _type = "Task" [ 1070.649993] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.657550] env[62965]: DEBUG oslo_vmware.api [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Task: {'id': task-1390367, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.701635] env[62965]: DEBUG nova.compute.manager [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1070.726887] env[62965]: DEBUG nova.virt.hardware [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1070.727246] env[62965]: DEBUG nova.virt.hardware [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1070.727419] env[62965]: DEBUG nova.virt.hardware [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1070.727610] env[62965]: DEBUG nova.virt.hardware [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1070.727764] env[62965]: DEBUG nova.virt.hardware [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1070.727918] env[62965]: DEBUG nova.virt.hardware [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1070.728152] env[62965]: DEBUG nova.virt.hardware [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1070.728320] env[62965]: DEBUG nova.virt.hardware [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1070.728499] env[62965]: DEBUG nova.virt.hardware [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1070.728664] env[62965]: DEBUG nova.virt.hardware [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1070.728844] env[62965]: DEBUG nova.virt.hardware [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1070.729821] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c8651e7-2c03-449d-90ea-a7b1acaafe4f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.737682] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-394a02a1-bddd-47b5-ac04-de75a82dd47a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.845697] env[62965]: DEBUG oslo_concurrency.lockutils [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.168s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.846526] env[62965]: DEBUG nova.compute.manager [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1070.940386] env[62965]: DEBUG nova.compute.manager [req-c05823bb-c0d8-4800-b50a-c91099955b82 req-1a95e0ff-06e4-47f1-b857-ba89b9a72f18 service nova] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Received event network-vif-plugged-5232b21c-f849-45de-916a-788114415557 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1070.940571] env[62965]: DEBUG oslo_concurrency.lockutils [req-c05823bb-c0d8-4800-b50a-c91099955b82 req-1a95e0ff-06e4-47f1-b857-ba89b9a72f18 service nova] Acquiring lock "b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.940786] env[62965]: DEBUG oslo_concurrency.lockutils [req-c05823bb-c0d8-4800-b50a-c91099955b82 req-1a95e0ff-06e4-47f1-b857-ba89b9a72f18 service nova] Lock "b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1070.940963] env[62965]: DEBUG oslo_concurrency.lockutils [req-c05823bb-c0d8-4800-b50a-c91099955b82 req-1a95e0ff-06e4-47f1-b857-ba89b9a72f18 service nova] Lock "b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.941149] env[62965]: DEBUG nova.compute.manager [req-c05823bb-c0d8-4800-b50a-c91099955b82 req-1a95e0ff-06e4-47f1-b857-ba89b9a72f18 service nova] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] No waiting events found dispatching network-vif-plugged-5232b21c-f849-45de-916a-788114415557 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1070.941320] env[62965]: WARNING nova.compute.manager [req-c05823bb-c0d8-4800-b50a-c91099955b82 req-1a95e0ff-06e4-47f1-b857-ba89b9a72f18 service nova] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Received unexpected event network-vif-plugged-5232b21c-f849-45de-916a-788114415557 for instance with vm_state building and task_state spawning. [ 1071.033598] env[62965]: DEBUG nova.network.neutron [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Successfully updated port: 5232b21c-f849-45de-916a-788114415557 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1071.100694] env[62965]: DEBUG oslo_concurrency.lockutils [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.100970] env[62965]: DEBUG oslo_concurrency.lockutils [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.102525] env[62965]: INFO nova.compute.claims [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1071.162180] env[62965]: DEBUG oslo_vmware.api [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Task: {'id': task-1390367, 'name': PowerOnVM_Task, 'duration_secs': 0.468135} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.162411] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1071.162917] env[62965]: INFO nova.compute.manager [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Took 7.15 seconds to spawn the instance on the hypervisor. [ 1071.162917] env[62965]: DEBUG nova.compute.manager [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1071.163598] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29c6f232-017c-4f4d-87f2-f622dcbffaff {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.352998] env[62965]: DEBUG nova.compute.utils [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1071.354420] env[62965]: DEBUG nova.compute.manager [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1071.354594] env[62965]: DEBUG nova.network.neutron [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1071.393187] env[62965]: DEBUG nova.policy [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e9d9f79c936143f5815fcb99755bbd4e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8fab418e2fe1420793517663574b43bb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 1071.536478] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "refresh_cache-b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1071.536879] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquired lock "refresh_cache-b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.536879] env[62965]: DEBUG nova.network.neutron [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1071.676099] env[62965]: DEBUG nova.network.neutron [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Successfully created port: 7de6b63c-f2c3-4a0b-b399-73075c5e09a5 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1071.686939] env[62965]: INFO nova.compute.manager [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Took 16.16 seconds to build instance. [ 1071.857793] env[62965]: DEBUG nova.compute.manager [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1072.071216] env[62965]: DEBUG nova.network.neutron [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1072.189928] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ac1f7585-d873-4395-b5ed-3c4bb64e2a47 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Lock "740b4d83-afc5-4ee2-9130-8c1d0697e5e5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.674s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.210209] env[62965]: DEBUG nova.network.neutron [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Updating instance_info_cache with network_info: [{"id": "5232b21c-f849-45de-916a-788114415557", "address": "fa:16:3e:51:9d:68", "network": {"id": "62052700-1781-4933-93c8-fa1aafdb0fd2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1931798228-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "978666785b0743bb843de96585f96fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5232b21c-f8", "ovs_interfaceid": "5232b21c-f849-45de-916a-788114415557", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1072.214263] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c7a4a35-01f6-4e23-be9f-36813f197888 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.222611] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fe77bc0-d961-4fc2-80fd-e20a29400b95 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.254906] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa1c8438-d6a1-4cce-ab07-e63816c757c0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.262188] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bdf4cb8-bc68-4cb0-a69f-ee343b937f6d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.277224] env[62965]: DEBUG nova.compute.provider_tree [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1072.714800] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Releasing lock "refresh_cache-b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1072.715152] env[62965]: DEBUG nova.compute.manager [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Instance network_info: |[{"id": "5232b21c-f849-45de-916a-788114415557", "address": "fa:16:3e:51:9d:68", "network": {"id": "62052700-1781-4933-93c8-fa1aafdb0fd2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1931798228-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "978666785b0743bb843de96585f96fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5232b21c-f8", "ovs_interfaceid": "5232b21c-f849-45de-916a-788114415557", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1072.715596] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:51:9d:68', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a10c88d7-d13f-44fd-acee-7a734eb5f56a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5232b21c-f849-45de-916a-788114415557', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1072.723419] env[62965]: DEBUG oslo.service.loopingcall [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1072.723748] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1072.724169] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ce904735-9dc4-4390-a798-072d4011c0b9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.745405] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1072.745405] env[62965]: value = "task-1390368" [ 1072.745405] env[62965]: _type = "Task" [ 1072.745405] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.753845] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390368, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.780220] env[62965]: DEBUG nova.scheduler.client.report [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1072.867090] env[62965]: DEBUG nova.compute.manager [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1072.892762] env[62965]: DEBUG nova.virt.hardware [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1072.893047] env[62965]: DEBUG nova.virt.hardware [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1072.893220] env[62965]: DEBUG nova.virt.hardware [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1072.893408] env[62965]: DEBUG nova.virt.hardware [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1072.893559] env[62965]: DEBUG nova.virt.hardware [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1072.893837] env[62965]: DEBUG nova.virt.hardware [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1072.893937] env[62965]: DEBUG nova.virt.hardware [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1072.894112] env[62965]: DEBUG nova.virt.hardware [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1072.894286] env[62965]: DEBUG nova.virt.hardware [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1072.894452] env[62965]: DEBUG nova.virt.hardware [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1072.894628] env[62965]: DEBUG nova.virt.hardware [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1072.895556] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6a4598b-66b1-47b6-bf90-9af136e6eae2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.903860] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c616ab9-f5f4-499c-87b1-11f268a84788 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.975452] env[62965]: DEBUG nova.compute.manager [req-53c8d045-3d2d-480b-8633-a10b4c48cd11 req-facac1bf-5b28-4a21-bbfb-b50eca16357b service nova] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Received event network-changed-5232b21c-f849-45de-916a-788114415557 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1072.975659] env[62965]: DEBUG nova.compute.manager [req-53c8d045-3d2d-480b-8633-a10b4c48cd11 req-facac1bf-5b28-4a21-bbfb-b50eca16357b service nova] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Refreshing instance network info cache due to event network-changed-5232b21c-f849-45de-916a-788114415557. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 1072.975879] env[62965]: DEBUG oslo_concurrency.lockutils [req-53c8d045-3d2d-480b-8633-a10b4c48cd11 req-facac1bf-5b28-4a21-bbfb-b50eca16357b service nova] Acquiring lock "refresh_cache-b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1072.976028] env[62965]: DEBUG oslo_concurrency.lockutils [req-53c8d045-3d2d-480b-8633-a10b4c48cd11 req-facac1bf-5b28-4a21-bbfb-b50eca16357b service nova] Acquired lock "refresh_cache-b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.976198] env[62965]: DEBUG nova.network.neutron [req-53c8d045-3d2d-480b-8633-a10b4c48cd11 req-facac1bf-5b28-4a21-bbfb-b50eca16357b service nova] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Refreshing network info cache for port 5232b21c-f849-45de-916a-788114415557 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1072.979034] env[62965]: DEBUG oslo_concurrency.lockutils [None req-505a420d-1c7f-4bb4-9a41-effacde01b53 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Acquiring lock "740b4d83-afc5-4ee2-9130-8c1d0697e5e5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.979291] env[62965]: DEBUG oslo_concurrency.lockutils [None req-505a420d-1c7f-4bb4-9a41-effacde01b53 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Lock "740b4d83-afc5-4ee2-9130-8c1d0697e5e5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.979459] env[62965]: DEBUG oslo_concurrency.lockutils [None req-505a420d-1c7f-4bb4-9a41-effacde01b53 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Acquiring lock "740b4d83-afc5-4ee2-9130-8c1d0697e5e5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.979661] env[62965]: DEBUG oslo_concurrency.lockutils [None req-505a420d-1c7f-4bb4-9a41-effacde01b53 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Lock "740b4d83-afc5-4ee2-9130-8c1d0697e5e5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.979862] env[62965]: DEBUG oslo_concurrency.lockutils [None req-505a420d-1c7f-4bb4-9a41-effacde01b53 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Lock "740b4d83-afc5-4ee2-9130-8c1d0697e5e5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.982130] env[62965]: INFO nova.compute.manager [None req-505a420d-1c7f-4bb4-9a41-effacde01b53 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Terminating instance [ 1073.098209] env[62965]: DEBUG nova.compute.manager [req-2d4d239f-ef0f-4ab3-89f1-9cb5e597c54d req-70b9e991-1351-46f7-ae05-f98f7ef55d2a service nova] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Received event network-vif-plugged-7de6b63c-f2c3-4a0b-b399-73075c5e09a5 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1073.098469] env[62965]: DEBUG oslo_concurrency.lockutils [req-2d4d239f-ef0f-4ab3-89f1-9cb5e597c54d req-70b9e991-1351-46f7-ae05-f98f7ef55d2a service nova] Acquiring lock "385648d7-20e3-442c-b6a1-3d972d266487-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.098710] env[62965]: DEBUG oslo_concurrency.lockutils [req-2d4d239f-ef0f-4ab3-89f1-9cb5e597c54d req-70b9e991-1351-46f7-ae05-f98f7ef55d2a service nova] Lock "385648d7-20e3-442c-b6a1-3d972d266487-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.098993] env[62965]: DEBUG oslo_concurrency.lockutils [req-2d4d239f-ef0f-4ab3-89f1-9cb5e597c54d req-70b9e991-1351-46f7-ae05-f98f7ef55d2a service nova] Lock "385648d7-20e3-442c-b6a1-3d972d266487-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.099088] env[62965]: DEBUG nova.compute.manager [req-2d4d239f-ef0f-4ab3-89f1-9cb5e597c54d req-70b9e991-1351-46f7-ae05-f98f7ef55d2a service nova] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] No waiting events found dispatching network-vif-plugged-7de6b63c-f2c3-4a0b-b399-73075c5e09a5 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1073.099281] env[62965]: WARNING nova.compute.manager [req-2d4d239f-ef0f-4ab3-89f1-9cb5e597c54d req-70b9e991-1351-46f7-ae05-f98f7ef55d2a service nova] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Received unexpected event network-vif-plugged-7de6b63c-f2c3-4a0b-b399-73075c5e09a5 for instance with vm_state building and task_state spawning. [ 1073.255132] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390368, 'name': CreateVM_Task, 'duration_secs': 0.411924} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.255288] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1073.255920] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1073.256108] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.256442] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1073.256692] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc36c42f-9769-46c1-94b3-8afddceaac3b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.261593] env[62965]: DEBUG oslo_vmware.api [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 1073.261593] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5212c0dd-0f9a-b041-822e-11caa9befa46" [ 1073.261593] env[62965]: _type = "Task" [ 1073.261593] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.269153] env[62965]: DEBUG oslo_vmware.api [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5212c0dd-0f9a-b041-822e-11caa9befa46, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.284917] env[62965]: DEBUG oslo_concurrency.lockutils [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.184s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.285400] env[62965]: DEBUG nova.compute.manager [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1073.486492] env[62965]: DEBUG nova.compute.manager [None req-505a420d-1c7f-4bb4-9a41-effacde01b53 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1073.486751] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-505a420d-1c7f-4bb4-9a41-effacde01b53 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1073.487750] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70c79060-5bb6-4068-b5f0-c4956dc04ac5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.495256] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-505a420d-1c7f-4bb4-9a41-effacde01b53 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1073.495496] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-165c58d0-06cf-4ef1-bba4-04e6dac51bb7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.502749] env[62965]: DEBUG oslo_vmware.api [None req-505a420d-1c7f-4bb4-9a41-effacde01b53 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Waiting for the task: (returnval){ [ 1073.502749] env[62965]: value = "task-1390369" [ 1073.502749] env[62965]: _type = "Task" [ 1073.502749] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.511383] env[62965]: DEBUG oslo_vmware.api [None req-505a420d-1c7f-4bb4-9a41-effacde01b53 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Task: {'id': task-1390369, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.728261] env[62965]: DEBUG nova.network.neutron [req-53c8d045-3d2d-480b-8633-a10b4c48cd11 req-facac1bf-5b28-4a21-bbfb-b50eca16357b service nova] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Updated VIF entry in instance network info cache for port 5232b21c-f849-45de-916a-788114415557. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1073.728653] env[62965]: DEBUG nova.network.neutron [req-53c8d045-3d2d-480b-8633-a10b4c48cd11 req-facac1bf-5b28-4a21-bbfb-b50eca16357b service nova] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Updating instance_info_cache with network_info: [{"id": "5232b21c-f849-45de-916a-788114415557", "address": "fa:16:3e:51:9d:68", "network": {"id": "62052700-1781-4933-93c8-fa1aafdb0fd2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1931798228-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "978666785b0743bb843de96585f96fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5232b21c-f8", "ovs_interfaceid": "5232b21c-f849-45de-916a-788114415557", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1073.775987] env[62965]: DEBUG oslo_vmware.api [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5212c0dd-0f9a-b041-822e-11caa9befa46, 'name': SearchDatastore_Task, 'duration_secs': 0.009251} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.776651] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1073.776897] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1073.777164] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1073.777518] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1073.777629] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1073.777825] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0899a972-8cbc-4c1c-9b6b-b7530ce0591c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.785979] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1073.786191] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1073.786907] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff7ecd9e-9d97-42b3-bc25-4f149f875670 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.790729] env[62965]: DEBUG nova.compute.utils [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1073.792069] env[62965]: DEBUG nova.compute.manager [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Not allocating networking since 'none' was specified. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 1073.797410] env[62965]: DEBUG oslo_vmware.api [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 1073.797410] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]527389be-346b-404e-842e-8029fcf77a05" [ 1073.797410] env[62965]: _type = "Task" [ 1073.797410] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.807779] env[62965]: DEBUG oslo_vmware.api [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]527389be-346b-404e-842e-8029fcf77a05, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.848805] env[62965]: DEBUG nova.network.neutron [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Successfully updated port: 7de6b63c-f2c3-4a0b-b399-73075c5e09a5 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1074.013381] env[62965]: DEBUG oslo_vmware.api [None req-505a420d-1c7f-4bb4-9a41-effacde01b53 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Task: {'id': task-1390369, 'name': PowerOffVM_Task, 'duration_secs': 0.217371} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.013597] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-505a420d-1c7f-4bb4-9a41-effacde01b53 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1074.013824] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-505a420d-1c7f-4bb4-9a41-effacde01b53 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1074.014140] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-68042198-3293-40cb-ac6d-819124de9c56 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.094129] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-505a420d-1c7f-4bb4-9a41-effacde01b53 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1074.094595] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-505a420d-1c7f-4bb4-9a41-effacde01b53 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Deleting contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1074.094961] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-505a420d-1c7f-4bb4-9a41-effacde01b53 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Deleting the datastore file [datastore1] 740b4d83-afc5-4ee2-9130-8c1d0697e5e5 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1074.095408] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6fa88355-d30a-4aea-859c-32e4f6a4fcce {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.104414] env[62965]: DEBUG oslo_vmware.api [None req-505a420d-1c7f-4bb4-9a41-effacde01b53 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Waiting for the task: (returnval){ [ 1074.104414] env[62965]: value = "task-1390371" [ 1074.104414] env[62965]: _type = "Task" [ 1074.104414] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.112836] env[62965]: DEBUG oslo_vmware.api [None req-505a420d-1c7f-4bb4-9a41-effacde01b53 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Task: {'id': task-1390371, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.231610] env[62965]: DEBUG oslo_concurrency.lockutils [req-53c8d045-3d2d-480b-8633-a10b4c48cd11 req-facac1bf-5b28-4a21-bbfb-b50eca16357b service nova] Releasing lock "refresh_cache-b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1074.294831] env[62965]: DEBUG nova.compute.manager [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1074.307495] env[62965]: DEBUG oslo_vmware.api [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]527389be-346b-404e-842e-8029fcf77a05, 'name': SearchDatastore_Task, 'duration_secs': 0.009659} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.308413] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff2590ea-53fd-45d1-a94c-fa173e1818dc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.314240] env[62965]: DEBUG oslo_vmware.api [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 1074.314240] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]520d30b9-6419-b990-6aa7-63f69e58a718" [ 1074.314240] env[62965]: _type = "Task" [ 1074.314240] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.321987] env[62965]: DEBUG oslo_vmware.api [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]520d30b9-6419-b990-6aa7-63f69e58a718, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.351463] env[62965]: DEBUG oslo_concurrency.lockutils [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "refresh_cache-385648d7-20e3-442c-b6a1-3d972d266487" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1074.351612] env[62965]: DEBUG oslo_concurrency.lockutils [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquired lock "refresh_cache-385648d7-20e3-442c-b6a1-3d972d266487" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1074.351772] env[62965]: DEBUG nova.network.neutron [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1074.615485] env[62965]: DEBUG oslo_vmware.api [None req-505a420d-1c7f-4bb4-9a41-effacde01b53 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Task: {'id': task-1390371, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.168041} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.615691] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-505a420d-1c7f-4bb4-9a41-effacde01b53 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1074.615875] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-505a420d-1c7f-4bb4-9a41-effacde01b53 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Deleted contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1074.616068] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-505a420d-1c7f-4bb4-9a41-effacde01b53 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1074.616249] env[62965]: INFO nova.compute.manager [None req-505a420d-1c7f-4bb4-9a41-effacde01b53 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1074.616488] env[62965]: DEBUG oslo.service.loopingcall [None req-505a420d-1c7f-4bb4-9a41-effacde01b53 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1074.616970] env[62965]: DEBUG nova.compute.manager [-] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1074.617085] env[62965]: DEBUG nova.network.neutron [-] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1074.827204] env[62965]: DEBUG oslo_vmware.api [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]520d30b9-6419-b990-6aa7-63f69e58a718, 'name': SearchDatastore_Task, 'duration_secs': 0.009701} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.827445] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1074.827695] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7/b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1074.827952] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-009a10fa-38f5-4ffb-98f7-1b0ea91d69b6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.839020] env[62965]: DEBUG oslo_vmware.api [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 1074.839020] env[62965]: value = "task-1390372" [ 1074.839020] env[62965]: _type = "Task" [ 1074.839020] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.845792] env[62965]: DEBUG oslo_vmware.api [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390372, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.912019] env[62965]: DEBUG nova.network.neutron [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1075.004527] env[62965]: DEBUG nova.compute.manager [req-741ddb3c-dfaf-4070-82b7-40e82f292f49 req-53b1bf23-b8d7-469d-a582-53a362bb420d service nova] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Received event network-changed-7de6b63c-f2c3-4a0b-b399-73075c5e09a5 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1075.004755] env[62965]: DEBUG nova.compute.manager [req-741ddb3c-dfaf-4070-82b7-40e82f292f49 req-53b1bf23-b8d7-469d-a582-53a362bb420d service nova] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Refreshing instance network info cache due to event network-changed-7de6b63c-f2c3-4a0b-b399-73075c5e09a5. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 1075.005055] env[62965]: DEBUG oslo_concurrency.lockutils [req-741ddb3c-dfaf-4070-82b7-40e82f292f49 req-53b1bf23-b8d7-469d-a582-53a362bb420d service nova] Acquiring lock "refresh_cache-385648d7-20e3-442c-b6a1-3d972d266487" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1075.081945] env[62965]: DEBUG nova.network.neutron [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Updating instance_info_cache with network_info: [{"id": "7de6b63c-f2c3-4a0b-b399-73075c5e09a5", "address": "fa:16:3e:6c:49:73", "network": {"id": "4649fed7-c2ca-43a9-a183-7c99423da726", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1978698173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fab418e2fe1420793517663574b43bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43ad01d2-c7dd-453c-a929-8ad76294d13c", "external-id": "nsx-vlan-transportzone-176", "segmentation_id": 176, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7de6b63c-f2", "ovs_interfaceid": "7de6b63c-f2c3-4a0b-b399-73075c5e09a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1075.308465] env[62965]: DEBUG nova.compute.manager [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1075.335240] env[62965]: DEBUG nova.virt.hardware [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1075.335522] env[62965]: DEBUG nova.virt.hardware [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1075.335719] env[62965]: DEBUG nova.virt.hardware [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1075.335953] env[62965]: DEBUG nova.virt.hardware [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1075.336180] env[62965]: DEBUG nova.virt.hardware [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1075.336338] env[62965]: DEBUG nova.virt.hardware [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1075.336557] env[62965]: DEBUG nova.virt.hardware [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1075.336764] env[62965]: DEBUG nova.virt.hardware [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1075.336988] env[62965]: DEBUG nova.virt.hardware [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1075.337118] env[62965]: DEBUG nova.virt.hardware [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1075.337287] env[62965]: DEBUG nova.virt.hardware [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1075.338309] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5962299c-f7dc-4c0e-9be8-ce3cc4f6c044 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.350678] env[62965]: DEBUG oslo_vmware.api [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390372, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.354191] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10d91b45-c900-4932-bfc6-3c7cb219e30d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.358274] env[62965]: DEBUG nova.network.neutron [-] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1075.368955] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Instance VIF info [] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1075.374623] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Creating folder: Project (b5b2fd4a7b834ad0a27ac9bc6ac205d6). Parent ref: group-v294931. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1075.375471] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-76fb334c-94f0-45d7-adf1-f4acd4bcd3cc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.390419] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Created folder: Project (b5b2fd4a7b834ad0a27ac9bc6ac205d6) in parent group-v294931. [ 1075.390609] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Creating folder: Instances. Parent ref: group-v295059. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1075.390839] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-508bbafa-4eed-4043-aadf-7f873dd6f8d4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.403038] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Created folder: Instances in parent group-v295059. [ 1075.403038] env[62965]: DEBUG oslo.service.loopingcall [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1075.403038] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1075.403038] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f552c03f-a2e9-4761-9c5f-73fc1961602d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.418983] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1075.418983] env[62965]: value = "task-1390375" [ 1075.418983] env[62965]: _type = "Task" [ 1075.418983] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.426184] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390375, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.584599] env[62965]: DEBUG oslo_concurrency.lockutils [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Releasing lock "refresh_cache-385648d7-20e3-442c-b6a1-3d972d266487" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1075.584994] env[62965]: DEBUG nova.compute.manager [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Instance network_info: |[{"id": "7de6b63c-f2c3-4a0b-b399-73075c5e09a5", "address": "fa:16:3e:6c:49:73", "network": {"id": "4649fed7-c2ca-43a9-a183-7c99423da726", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1978698173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fab418e2fe1420793517663574b43bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43ad01d2-c7dd-453c-a929-8ad76294d13c", "external-id": "nsx-vlan-transportzone-176", "segmentation_id": 176, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7de6b63c-f2", "ovs_interfaceid": "7de6b63c-f2c3-4a0b-b399-73075c5e09a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1075.585395] env[62965]: DEBUG oslo_concurrency.lockutils [req-741ddb3c-dfaf-4070-82b7-40e82f292f49 req-53b1bf23-b8d7-469d-a582-53a362bb420d service nova] Acquired lock "refresh_cache-385648d7-20e3-442c-b6a1-3d972d266487" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.585635] env[62965]: DEBUG nova.network.neutron [req-741ddb3c-dfaf-4070-82b7-40e82f292f49 req-53b1bf23-b8d7-469d-a582-53a362bb420d service nova] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Refreshing network info cache for port 7de6b63c-f2c3-4a0b-b399-73075c5e09a5 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1075.587365] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6c:49:73', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '43ad01d2-c7dd-453c-a929-8ad76294d13c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7de6b63c-f2c3-4a0b-b399-73075c5e09a5', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1075.597692] env[62965]: DEBUG oslo.service.loopingcall [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1075.600707] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1075.601181] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d51235ed-28be-4ae4-9881-7108908c1c74 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.621900] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1075.621900] env[62965]: value = "task-1390376" [ 1075.621900] env[62965]: _type = "Task" [ 1075.621900] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.631700] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390376, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.818502] env[62965]: DEBUG nova.network.neutron [req-741ddb3c-dfaf-4070-82b7-40e82f292f49 req-53b1bf23-b8d7-469d-a582-53a362bb420d service nova] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Updated VIF entry in instance network info cache for port 7de6b63c-f2c3-4a0b-b399-73075c5e09a5. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1075.818867] env[62965]: DEBUG nova.network.neutron [req-741ddb3c-dfaf-4070-82b7-40e82f292f49 req-53b1bf23-b8d7-469d-a582-53a362bb420d service nova] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Updating instance_info_cache with network_info: [{"id": "7de6b63c-f2c3-4a0b-b399-73075c5e09a5", "address": "fa:16:3e:6c:49:73", "network": {"id": "4649fed7-c2ca-43a9-a183-7c99423da726", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1978698173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fab418e2fe1420793517663574b43bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43ad01d2-c7dd-453c-a929-8ad76294d13c", "external-id": "nsx-vlan-transportzone-176", "segmentation_id": 176, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7de6b63c-f2", "ovs_interfaceid": "7de6b63c-f2c3-4a0b-b399-73075c5e09a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1075.847765] env[62965]: DEBUG oslo_vmware.api [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390372, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.552188} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.847765] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7/b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1075.847765] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1075.847765] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2127202b-efc4-418e-a52b-4caffde24487 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.854819] env[62965]: DEBUG oslo_vmware.api [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 1075.854819] env[62965]: value = "task-1390377" [ 1075.854819] env[62965]: _type = "Task" [ 1075.854819] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.862975] env[62965]: INFO nova.compute.manager [-] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Took 1.25 seconds to deallocate network for instance. [ 1075.863263] env[62965]: DEBUG oslo_vmware.api [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390377, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.929269] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390375, 'name': CreateVM_Task, 'duration_secs': 0.288657} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.929440] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1075.929877] env[62965]: DEBUG oslo_concurrency.lockutils [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1075.930060] env[62965]: DEBUG oslo_concurrency.lockutils [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1075.930388] env[62965]: DEBUG oslo_concurrency.lockutils [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1075.930641] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-15738f07-ce35-4b72-867c-e278942af137 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.934922] env[62965]: DEBUG oslo_vmware.api [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Waiting for the task: (returnval){ [ 1075.934922] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52433fe8-5332-1526-0756-25ebdc8d4c18" [ 1075.934922] env[62965]: _type = "Task" [ 1075.934922] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.942144] env[62965]: DEBUG oslo_vmware.api [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52433fe8-5332-1526-0756-25ebdc8d4c18, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.132454] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390376, 'name': CreateVM_Task} progress is 99%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.322385] env[62965]: DEBUG oslo_concurrency.lockutils [req-741ddb3c-dfaf-4070-82b7-40e82f292f49 req-53b1bf23-b8d7-469d-a582-53a362bb420d service nova] Releasing lock "refresh_cache-385648d7-20e3-442c-b6a1-3d972d266487" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1076.322385] env[62965]: DEBUG nova.compute.manager [req-741ddb3c-dfaf-4070-82b7-40e82f292f49 req-53b1bf23-b8d7-469d-a582-53a362bb420d service nova] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Received event network-vif-deleted-19aed2fc-7ec6-4240-9c8e-a109f243d2f5 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1076.322385] env[62965]: INFO nova.compute.manager [req-741ddb3c-dfaf-4070-82b7-40e82f292f49 req-53b1bf23-b8d7-469d-a582-53a362bb420d service nova] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Neutron deleted interface 19aed2fc-7ec6-4240-9c8e-a109f243d2f5; detaching it from the instance and deleting it from the info cache [ 1076.322385] env[62965]: DEBUG nova.network.neutron [req-741ddb3c-dfaf-4070-82b7-40e82f292f49 req-53b1bf23-b8d7-469d-a582-53a362bb420d service nova] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1076.364503] env[62965]: DEBUG oslo_vmware.api [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390377, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.369398] env[62965]: DEBUG oslo_concurrency.lockutils [None req-505a420d-1c7f-4bb4-9a41-effacde01b53 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.369643] env[62965]: DEBUG oslo_concurrency.lockutils [None req-505a420d-1c7f-4bb4-9a41-effacde01b53 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.369855] env[62965]: DEBUG nova.objects.instance [None req-505a420d-1c7f-4bb4-9a41-effacde01b53 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Lazy-loading 'resources' on Instance uuid 740b4d83-afc5-4ee2-9130-8c1d0697e5e5 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1076.444773] env[62965]: DEBUG oslo_vmware.api [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52433fe8-5332-1526-0756-25ebdc8d4c18, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.632698] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390376, 'name': CreateVM_Task, 'duration_secs': 0.564107} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.632859] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1076.633504] env[62965]: DEBUG oslo_concurrency.lockutils [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1076.825300] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5ea71f7b-05c3-4d37-8b31-b12bb69d18b7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.834832] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9d1fa6a-7771-4225-b4f8-66c86442fd33 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.859636] env[62965]: DEBUG nova.compute.manager [req-741ddb3c-dfaf-4070-82b7-40e82f292f49 req-53b1bf23-b8d7-469d-a582-53a362bb420d service nova] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Detach interface failed, port_id=19aed2fc-7ec6-4240-9c8e-a109f243d2f5, reason: Instance 740b4d83-afc5-4ee2-9130-8c1d0697e5e5 could not be found. {{(pid=62965) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11294}} [ 1076.866666] env[62965]: DEBUG oslo_vmware.api [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390377, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.949173] env[62965]: DEBUG oslo_vmware.api [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52433fe8-5332-1526-0756-25ebdc8d4c18, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.953233] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5418a23a-b1ee-44d4-af21-a1a9b1aff08a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.959985] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbe4b982-b812-4cbe-8aab-3682f4bf22f5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.989860] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54945f26-fab5-424a-916e-f1ba1ef7be5d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.997167] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0e13c5d-e166-49c7-9c8c-c0545cb41837 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.010196] env[62965]: DEBUG nova.compute.provider_tree [None req-505a420d-1c7f-4bb4-9a41-effacde01b53 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1077.368734] env[62965]: DEBUG oslo_vmware.api [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390377, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.447387] env[62965]: DEBUG oslo_vmware.api [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52433fe8-5332-1526-0756-25ebdc8d4c18, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.513023] env[62965]: DEBUG nova.scheduler.client.report [None req-505a420d-1c7f-4bb4-9a41-effacde01b53 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1077.869719] env[62965]: DEBUG oslo_vmware.api [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390377, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.946649] env[62965]: DEBUG oslo_vmware.api [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52433fe8-5332-1526-0756-25ebdc8d4c18, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.018028] env[62965]: DEBUG oslo_concurrency.lockutils [None req-505a420d-1c7f-4bb4-9a41-effacde01b53 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.648s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.041048] env[62965]: INFO nova.scheduler.client.report [None req-505a420d-1c7f-4bb4-9a41-effacde01b53 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Deleted allocations for instance 740b4d83-afc5-4ee2-9130-8c1d0697e5e5 [ 1078.370340] env[62965]: DEBUG oslo_vmware.api [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390377, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.447167] env[62965]: DEBUG oslo_vmware.api [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52433fe8-5332-1526-0756-25ebdc8d4c18, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.548310] env[62965]: DEBUG oslo_concurrency.lockutils [None req-505a420d-1c7f-4bb4-9a41-effacde01b53 tempest-ServerGroupTestJSON-1901650150 tempest-ServerGroupTestJSON-1901650150-project-member] Lock "740b4d83-afc5-4ee2-9130-8c1d0697e5e5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.569s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.870595] env[62965]: DEBUG oslo_vmware.api [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390377, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.947694] env[62965]: DEBUG oslo_vmware.api [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52433fe8-5332-1526-0756-25ebdc8d4c18, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.373496] env[62965]: DEBUG oslo_vmware.api [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390377, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.449381] env[62965]: DEBUG oslo_vmware.api [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52433fe8-5332-1526-0756-25ebdc8d4c18, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.878295] env[62965]: DEBUG oslo_vmware.api [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390377, 'name': ExtendVirtualDisk_Task, 'duration_secs': 3.821209} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.878664] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1079.879484] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21506f27-57d9-48fa-b245-7971a1419bac {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.901051] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7/b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1079.901303] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c2038baf-14be-434c-b9cd-9822c2182afe {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.921795] env[62965]: DEBUG oslo_vmware.api [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 1079.921795] env[62965]: value = "task-1390378" [ 1079.921795] env[62965]: _type = "Task" [ 1079.921795] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.928119] env[62965]: DEBUG oslo_vmware.api [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390378, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.952027] env[62965]: DEBUG oslo_vmware.api [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52433fe8-5332-1526-0756-25ebdc8d4c18, 'name': SearchDatastore_Task, 'duration_secs': 3.53604} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.952592] env[62965]: DEBUG oslo_concurrency.lockutils [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1079.952837] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1079.953098] env[62965]: DEBUG oslo_concurrency.lockutils [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1079.953510] env[62965]: DEBUG oslo_concurrency.lockutils [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1079.953680] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1079.953971] env[62965]: DEBUG oslo_concurrency.lockutils [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1079.954311] env[62965]: DEBUG oslo_concurrency.lockutils [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1079.954715] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ec0fbbe0-b709-4b46-9312-7523b0885805 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.956693] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-663507df-e44e-49d7-ae48-33b8438a1ace {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.961604] env[62965]: DEBUG oslo_vmware.api [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1079.961604] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5252bd92-8caa-a540-6e64-f505c45d1314" [ 1079.961604] env[62965]: _type = "Task" [ 1079.961604] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.965534] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1079.965733] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1079.966759] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41a2c1db-6290-4a73-9ca8-d3c34e6d3b16 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.972273] env[62965]: DEBUG oslo_vmware.api [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5252bd92-8caa-a540-6e64-f505c45d1314, 'name': SearchDatastore_Task, 'duration_secs': 0.007895} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.973170] env[62965]: DEBUG oslo_concurrency.lockutils [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1079.973170] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1079.973311] env[62965]: DEBUG oslo_concurrency.lockutils [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1079.975086] env[62965]: DEBUG oslo_vmware.api [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Waiting for the task: (returnval){ [ 1079.975086] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52972650-71ba-2ae7-ecda-0736d6cf5d01" [ 1079.975086] env[62965]: _type = "Task" [ 1079.975086] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.981633] env[62965]: DEBUG oslo_vmware.api [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52972650-71ba-2ae7-ecda-0736d6cf5d01, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.430650] env[62965]: DEBUG oslo_vmware.api [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390378, 'name': ReconfigVM_Task, 'duration_secs': 0.252469} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.430650] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Reconfigured VM instance instance-00000069 to attach disk [datastore1] b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7/b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1080.431232] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bbc03c54-b8c1-4224-84e3-21729bbc2d5e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.437265] env[62965]: DEBUG oslo_vmware.api [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 1080.437265] env[62965]: value = "task-1390379" [ 1080.437265] env[62965]: _type = "Task" [ 1080.437265] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.449957] env[62965]: DEBUG oslo_vmware.api [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390379, 'name': Rename_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.484117] env[62965]: DEBUG oslo_vmware.api [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52972650-71ba-2ae7-ecda-0736d6cf5d01, 'name': SearchDatastore_Task, 'duration_secs': 0.007205} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.484888] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83e29038-c42c-43cc-bd81-a094af5e667c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.493121] env[62965]: DEBUG oslo_vmware.api [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Waiting for the task: (returnval){ [ 1080.493121] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]526926c1-fea4-15e4-52ae-778c050611c3" [ 1080.493121] env[62965]: _type = "Task" [ 1080.493121] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.500447] env[62965]: DEBUG oslo_vmware.api [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]526926c1-fea4-15e4-52ae-778c050611c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.947974] env[62965]: DEBUG oslo_vmware.api [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390379, 'name': Rename_Task, 'duration_secs': 0.174549} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.948383] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1080.948699] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aba12e4d-74b1-4945-a1e6-02ecedb4b3ee {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.955118] env[62965]: DEBUG oslo_vmware.api [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 1080.955118] env[62965]: value = "task-1390380" [ 1080.955118] env[62965]: _type = "Task" [ 1080.955118] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.962589] env[62965]: DEBUG oslo_vmware.api [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390380, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.002517] env[62965]: DEBUG oslo_vmware.api [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]526926c1-fea4-15e4-52ae-778c050611c3, 'name': SearchDatastore_Task, 'duration_secs': 0.02728} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.002803] env[62965]: DEBUG oslo_concurrency.lockutils [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1081.003094] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] 182e796f-2fbf-4304-9339-dc6e9b6b5839/182e796f-2fbf-4304-9339-dc6e9b6b5839.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1081.003636] env[62965]: DEBUG oslo_concurrency.lockutils [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.003636] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1081.003774] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-847a6c71-1bfb-41a3-9bbc-0c7ea72dd982 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.005731] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4b84b12b-13c9-476a-af0a-54e80f2d92ed {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.011895] env[62965]: DEBUG oslo_vmware.api [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Waiting for the task: (returnval){ [ 1081.011895] env[62965]: value = "task-1390381" [ 1081.011895] env[62965]: _type = "Task" [ 1081.011895] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.015455] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1081.015629] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1081.016568] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a954b04b-e4d9-43a9-9523-9ce13cb7c469 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.021273] env[62965]: DEBUG oslo_vmware.api [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': task-1390381, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.024264] env[62965]: DEBUG oslo_vmware.api [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1081.024264] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]529e6ef3-ef43-a0fb-34fc-9dad50d56e20" [ 1081.024264] env[62965]: _type = "Task" [ 1081.024264] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.030846] env[62965]: DEBUG oslo_vmware.api [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]529e6ef3-ef43-a0fb-34fc-9dad50d56e20, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.465700] env[62965]: DEBUG oslo_vmware.api [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390380, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.521865] env[62965]: DEBUG oslo_vmware.api [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': task-1390381, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.503702} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.522186] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] 182e796f-2fbf-4304-9339-dc6e9b6b5839/182e796f-2fbf-4304-9339-dc6e9b6b5839.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1081.522403] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1081.522668] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c548ab1e-4f2d-4e0e-b6bf-d1535cc4d289 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.532737] env[62965]: DEBUG oslo_vmware.api [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]529e6ef3-ef43-a0fb-34fc-9dad50d56e20, 'name': SearchDatastore_Task, 'duration_secs': 0.009992} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.534463] env[62965]: DEBUG oslo_vmware.api [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Waiting for the task: (returnval){ [ 1081.534463] env[62965]: value = "task-1390382" [ 1081.534463] env[62965]: _type = "Task" [ 1081.534463] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.534660] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-292d35d2-a6d2-4bec-a0ac-6456e6daf0f2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.542356] env[62965]: DEBUG oslo_vmware.api [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1081.542356] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5234bfeb-88d3-0172-afd1-37a67ae09ee1" [ 1081.542356] env[62965]: _type = "Task" [ 1081.542356] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.545235] env[62965]: DEBUG oslo_vmware.api [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': task-1390382, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.552420] env[62965]: DEBUG oslo_vmware.api [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5234bfeb-88d3-0172-afd1-37a67ae09ee1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.965986] env[62965]: DEBUG oslo_vmware.api [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390380, 'name': PowerOnVM_Task, 'duration_secs': 0.719901} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.966358] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1081.966480] env[62965]: INFO nova.compute.manager [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Took 11.26 seconds to spawn the instance on the hypervisor. [ 1081.966666] env[62965]: DEBUG nova.compute.manager [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1081.967424] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90ed2bc5-ccbf-4305-9046-94accb19d0e0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.047611] env[62965]: DEBUG oslo_vmware.api [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': task-1390382, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083167} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.051934] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1082.053007] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be79fa5a-ec49-4569-b378-506e12c4ad9a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.064452] env[62965]: DEBUG oslo_vmware.api [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5234bfeb-88d3-0172-afd1-37a67ae09ee1, 'name': SearchDatastore_Task, 'duration_secs': 0.009473} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.074799] env[62965]: DEBUG oslo_concurrency.lockutils [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1082.075172] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] 385648d7-20e3-442c-b6a1-3d972d266487/385648d7-20e3-442c-b6a1-3d972d266487.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1082.088727] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] 182e796f-2fbf-4304-9339-dc6e9b6b5839/182e796f-2fbf-4304-9339-dc6e9b6b5839.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1082.089053] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5b3caab1-39d3-4dba-9756-3907d0608b42 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.091741] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4af60375-710e-4b94-a626-c530ef4f2fbf {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.119141] env[62965]: DEBUG oslo_vmware.api [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1082.119141] env[62965]: value = "task-1390383" [ 1082.119141] env[62965]: _type = "Task" [ 1082.119141] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.120609] env[62965]: DEBUG oslo_vmware.api [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Waiting for the task: (returnval){ [ 1082.120609] env[62965]: value = "task-1390384" [ 1082.120609] env[62965]: _type = "Task" [ 1082.120609] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.134729] env[62965]: DEBUG oslo_vmware.api [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390383, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.138716] env[62965]: DEBUG oslo_vmware.api [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': task-1390384, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.483606] env[62965]: INFO nova.compute.manager [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Took 16.88 seconds to build instance. [ 1082.630958] env[62965]: DEBUG oslo_vmware.api [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390383, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.636189] env[62965]: DEBUG oslo_vmware.api [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': task-1390384, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.986306] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e6777972-017b-4c50-a357-eeb0afcbe083 tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.387s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.134951] env[62965]: DEBUG oslo_vmware.api [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': task-1390384, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.138439] env[62965]: DEBUG oslo_vmware.api [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390383, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.632641] env[62965]: DEBUG oslo_vmware.api [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390383, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.305784} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.633253] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] 385648d7-20e3-442c-b6a1-3d972d266487/385648d7-20e3-442c-b6a1-3d972d266487.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1083.633516] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1083.633744] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-869bb941-0a11-4f29-9c94-e034c7bcb353 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.638057] env[62965]: DEBUG oslo_vmware.api [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': task-1390384, 'name': ReconfigVM_Task, 'duration_secs': 1.375638} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.638599] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Reconfigured VM instance instance-0000006b to attach disk [datastore1] 182e796f-2fbf-4304-9339-dc6e9b6b5839/182e796f-2fbf-4304-9339-dc6e9b6b5839.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1083.639186] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d1558427-f542-4b0a-a981-a19a82e794ed {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.643639] env[62965]: DEBUG oslo_vmware.api [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1083.643639] env[62965]: value = "task-1390385" [ 1083.643639] env[62965]: _type = "Task" [ 1083.643639] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.644894] env[62965]: DEBUG oslo_vmware.api [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Waiting for the task: (returnval){ [ 1083.644894] env[62965]: value = "task-1390386" [ 1083.644894] env[62965]: _type = "Task" [ 1083.644894] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.657771] env[62965]: DEBUG oslo_vmware.api [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390385, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.661438] env[62965]: DEBUG oslo_vmware.api [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': task-1390386, 'name': Rename_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.070744] env[62965]: DEBUG nova.compute.manager [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Stashing vm_state: active {{(pid=62965) _prep_resize /opt/stack/nova/nova/compute/manager.py:5953}} [ 1084.158504] env[62965]: DEBUG oslo_vmware.api [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': task-1390386, 'name': Rename_Task, 'duration_secs': 0.151598} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.161192] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1084.161453] env[62965]: DEBUG oslo_vmware.api [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390385, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076437} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.161653] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-40158b20-7a73-4707-9c38-cce8b1329951 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.163073] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1084.163781] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c83078b-d1cb-4aff-b5f7-b0d1a29733cf {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.170748] env[62965]: DEBUG oslo_vmware.api [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Waiting for the task: (returnval){ [ 1084.170748] env[62965]: value = "task-1390387" [ 1084.170748] env[62965]: _type = "Task" [ 1084.170748] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.187934] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] 385648d7-20e3-442c-b6a1-3d972d266487/385648d7-20e3-442c-b6a1-3d972d266487.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1084.191059] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7980080a-7b22-48a4-a13b-1546130c88ef {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.210159] env[62965]: DEBUG oslo_vmware.api [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': task-1390387, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.211364] env[62965]: DEBUG oslo_vmware.api [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1084.211364] env[62965]: value = "task-1390388" [ 1084.211364] env[62965]: _type = "Task" [ 1084.211364] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.220074] env[62965]: DEBUG oslo_vmware.api [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390388, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.593623] env[62965]: DEBUG oslo_concurrency.lockutils [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.593902] env[62965]: DEBUG oslo_concurrency.lockutils [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.689462] env[62965]: DEBUG oslo_vmware.api [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': task-1390387, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.720397] env[62965]: DEBUG oslo_vmware.api [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390388, 'name': ReconfigVM_Task, 'duration_secs': 0.310689} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.720705] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Reconfigured VM instance instance-0000006a to attach disk [datastore1] 385648d7-20e3-442c-b6a1-3d972d266487/385648d7-20e3-442c-b6a1-3d972d266487.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1084.721406] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-66360d6f-06bc-4cab-9304-26ee6a3d4126 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.728744] env[62965]: DEBUG oslo_vmware.api [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1084.728744] env[62965]: value = "task-1390389" [ 1084.728744] env[62965]: _type = "Task" [ 1084.728744] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.737021] env[62965]: DEBUG oslo_vmware.api [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390389, 'name': Rename_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.100093] env[62965]: INFO nova.compute.claims [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1085.193630] env[62965]: DEBUG oslo_vmware.api [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': task-1390387, 'name': PowerOnVM_Task, 'duration_secs': 0.599868} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.194025] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1085.194336] env[62965]: INFO nova.compute.manager [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Took 9.89 seconds to spawn the instance on the hypervisor. [ 1085.194607] env[62965]: DEBUG nova.compute.manager [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1085.195685] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e80a5b6c-8a2d-4e51-ab63-a1679ec11ab5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.240055] env[62965]: DEBUG oslo_vmware.api [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390389, 'name': Rename_Task, 'duration_secs': 0.149191} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.240393] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1085.240705] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-82aad7cc-729f-420f-8b4e-995c6d27d55f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.247483] env[62965]: DEBUG oslo_vmware.api [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1085.247483] env[62965]: value = "task-1390390" [ 1085.247483] env[62965]: _type = "Task" [ 1085.247483] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.257197] env[62965]: DEBUG oslo_vmware.api [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390390, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.607612] env[62965]: INFO nova.compute.resource_tracker [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Updating resource usage from migration 330e662e-d3f2-4c7d-84ec-bf24984e877a [ 1085.691163] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e152ff5b-cf2a-48c8-a073-ec9d2101080c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.700271] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98ae3cb0-0be8-422b-97e6-83d3569168dc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.737031] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-043fb9d1-e1bd-4cc9-ac3f-f6d375d6f4cc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.742964] env[62965]: INFO nova.compute.manager [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Took 14.66 seconds to build instance. [ 1085.751754] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61dd10dc-361a-4bd2-b1b1-76d3d9d8fd2c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.770038] env[62965]: DEBUG nova.compute.provider_tree [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1085.771009] env[62965]: DEBUG oslo_vmware.api [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390390, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.244596] env[62965]: INFO nova.compute.manager [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Rebuilding instance [ 1086.247468] env[62965]: DEBUG oslo_concurrency.lockutils [None req-0d59547c-91cd-4fff-ad09-ccc9b777fdfe tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Lock "182e796f-2fbf-4304-9339-dc6e9b6b5839" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.179s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1086.258495] env[62965]: DEBUG oslo_vmware.api [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390390, 'name': PowerOnVM_Task, 'duration_secs': 0.911879} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.258763] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1086.258966] env[62965]: INFO nova.compute.manager [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Took 13.39 seconds to spawn the instance on the hypervisor. [ 1086.259164] env[62965]: DEBUG nova.compute.manager [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1086.259941] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4ad8419-c6ca-48df-9111-ee3cd9c0ef95 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.274463] env[62965]: DEBUG nova.scheduler.client.report [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1086.293409] env[62965]: DEBUG nova.compute.manager [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1086.294291] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fad9f55e-3116-41ef-aaff-64a99a17f22b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.783130] env[62965]: DEBUG oslo_concurrency.lockutils [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.189s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1086.783246] env[62965]: INFO nova.compute.manager [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Migrating [ 1086.793567] env[62965]: INFO nova.compute.manager [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Took 18.96 seconds to build instance. [ 1087.299656] env[62965]: DEBUG oslo_concurrency.lockutils [None req-216f2e41-18ac-4b65-a33c-d0f342263ad7 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "385648d7-20e3-442c-b6a1-3d972d266487" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.475s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.300134] env[62965]: DEBUG oslo_concurrency.lockutils [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "refresh_cache-b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1087.300301] env[62965]: DEBUG oslo_concurrency.lockutils [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquired lock "refresh_cache-b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.300480] env[62965]: DEBUG nova.network.neutron [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1087.309451] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1087.310048] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-282c0aff-1668-423e-b541-2256489eee2a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.318342] env[62965]: DEBUG oslo_vmware.api [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Waiting for the task: (returnval){ [ 1087.318342] env[62965]: value = "task-1390391" [ 1087.318342] env[62965]: _type = "Task" [ 1087.318342] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.327932] env[62965]: DEBUG oslo_vmware.api [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': task-1390391, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.831584] env[62965]: DEBUG oslo_vmware.api [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': task-1390391, 'name': PowerOffVM_Task, 'duration_secs': 0.254357} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.831584] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1087.831584] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1087.832392] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8533b4f0-658a-455f-ac1b-23a2c25a9492 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.841033] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1087.841033] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0edff060-11db-4dc4-9da3-1195ce5e4da6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.869758] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1087.869957] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Deleting contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1087.870168] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Deleting the datastore file [datastore1] 182e796f-2fbf-4304-9339-dc6e9b6b5839 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1087.870441] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-526cb84c-5e4e-4ad4-bca2-0605f04eb015 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.877725] env[62965]: DEBUG oslo_vmware.api [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Waiting for the task: (returnval){ [ 1087.877725] env[62965]: value = "task-1390393" [ 1087.877725] env[62965]: _type = "Task" [ 1087.877725] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.889922] env[62965]: DEBUG oslo_vmware.api [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': task-1390393, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.011124] env[62965]: DEBUG nova.compute.manager [req-f6c40fd8-5e6b-4b96-b99f-8d99fcfbc7fd req-7590fcf0-6617-4dc6-bf0d-6eb9168c7b4a service nova] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Received event network-changed-7de6b63c-f2c3-4a0b-b399-73075c5e09a5 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1088.011334] env[62965]: DEBUG nova.compute.manager [req-f6c40fd8-5e6b-4b96-b99f-8d99fcfbc7fd req-7590fcf0-6617-4dc6-bf0d-6eb9168c7b4a service nova] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Refreshing instance network info cache due to event network-changed-7de6b63c-f2c3-4a0b-b399-73075c5e09a5. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 1088.011549] env[62965]: DEBUG oslo_concurrency.lockutils [req-f6c40fd8-5e6b-4b96-b99f-8d99fcfbc7fd req-7590fcf0-6617-4dc6-bf0d-6eb9168c7b4a service nova] Acquiring lock "refresh_cache-385648d7-20e3-442c-b6a1-3d972d266487" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1088.011690] env[62965]: DEBUG oslo_concurrency.lockutils [req-f6c40fd8-5e6b-4b96-b99f-8d99fcfbc7fd req-7590fcf0-6617-4dc6-bf0d-6eb9168c7b4a service nova] Acquired lock "refresh_cache-385648d7-20e3-442c-b6a1-3d972d266487" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1088.014023] env[62965]: DEBUG nova.network.neutron [req-f6c40fd8-5e6b-4b96-b99f-8d99fcfbc7fd req-7590fcf0-6617-4dc6-bf0d-6eb9168c7b4a service nova] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Refreshing network info cache for port 7de6b63c-f2c3-4a0b-b399-73075c5e09a5 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1088.085947] env[62965]: DEBUG nova.network.neutron [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Updating instance_info_cache with network_info: [{"id": "5232b21c-f849-45de-916a-788114415557", "address": "fa:16:3e:51:9d:68", "network": {"id": "62052700-1781-4933-93c8-fa1aafdb0fd2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1931798228-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "978666785b0743bb843de96585f96fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5232b21c-f8", "ovs_interfaceid": "5232b21c-f849-45de-916a-788114415557", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1088.387888] env[62965]: DEBUG oslo_vmware.api [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': task-1390393, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.229271} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.388237] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1088.388334] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Deleted contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1088.388514] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1088.589036] env[62965]: DEBUG oslo_concurrency.lockutils [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Releasing lock "refresh_cache-b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1088.758944] env[62965]: DEBUG nova.network.neutron [req-f6c40fd8-5e6b-4b96-b99f-8d99fcfbc7fd req-7590fcf0-6617-4dc6-bf0d-6eb9168c7b4a service nova] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Updated VIF entry in instance network info cache for port 7de6b63c-f2c3-4a0b-b399-73075c5e09a5. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1088.759414] env[62965]: DEBUG nova.network.neutron [req-f6c40fd8-5e6b-4b96-b99f-8d99fcfbc7fd req-7590fcf0-6617-4dc6-bf0d-6eb9168c7b4a service nova] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Updating instance_info_cache with network_info: [{"id": "7de6b63c-f2c3-4a0b-b399-73075c5e09a5", "address": "fa:16:3e:6c:49:73", "network": {"id": "4649fed7-c2ca-43a9-a183-7c99423da726", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1978698173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fab418e2fe1420793517663574b43bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43ad01d2-c7dd-453c-a929-8ad76294d13c", "external-id": "nsx-vlan-transportzone-176", "segmentation_id": 176, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7de6b63c-f2", "ovs_interfaceid": "7de6b63c-f2c3-4a0b-b399-73075c5e09a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1089.262327] env[62965]: DEBUG oslo_concurrency.lockutils [req-f6c40fd8-5e6b-4b96-b99f-8d99fcfbc7fd req-7590fcf0-6617-4dc6-bf0d-6eb9168c7b4a service nova] Releasing lock "refresh_cache-385648d7-20e3-442c-b6a1-3d972d266487" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1089.423943] env[62965]: DEBUG nova.virt.hardware [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1089.424264] env[62965]: DEBUG nova.virt.hardware [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1089.424403] env[62965]: DEBUG nova.virt.hardware [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1089.424592] env[62965]: DEBUG nova.virt.hardware [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1089.424742] env[62965]: DEBUG nova.virt.hardware [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1089.424891] env[62965]: DEBUG nova.virt.hardware [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1089.425119] env[62965]: DEBUG nova.virt.hardware [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1089.425285] env[62965]: DEBUG nova.virt.hardware [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1089.425459] env[62965]: DEBUG nova.virt.hardware [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1089.425623] env[62965]: DEBUG nova.virt.hardware [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1089.425802] env[62965]: DEBUG nova.virt.hardware [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1089.426691] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b57507b-d409-40ee-8cad-362f66e3a9e6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.436420] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e326924f-9a11-4f74-887f-bb4997785991 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.451707] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Instance VIF info [] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1089.458055] env[62965]: DEBUG oslo.service.loopingcall [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1089.458055] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1089.458290] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c3e6a55f-4f0a-43b6-b378-9db46a5e0cf9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.475476] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1089.475476] env[62965]: value = "task-1390394" [ 1089.475476] env[62965]: _type = "Task" [ 1089.475476] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.483364] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390394, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.985859] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390394, 'name': CreateVM_Task} progress is 99%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.105591] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17bcc79e-2794-42d2-9880-d255a96fc368 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.124526] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Updating instance 'b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7' progress to 0 {{(pid=62965) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1090.491936] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390394, 'name': CreateVM_Task, 'duration_secs': 0.661407} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.492373] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1090.492743] env[62965]: DEBUG oslo_concurrency.lockutils [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1090.493057] env[62965]: DEBUG oslo_concurrency.lockutils [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1090.493564] env[62965]: DEBUG oslo_concurrency.lockutils [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1090.493922] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb71e055-a7ba-4f85-a1d5-087929a82684 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.499432] env[62965]: DEBUG oslo_vmware.api [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Waiting for the task: (returnval){ [ 1090.499432] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52a37d1f-9e1d-0c11-8690-8a0590eb00dc" [ 1090.499432] env[62965]: _type = "Task" [ 1090.499432] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.509842] env[62965]: DEBUG oslo_vmware.api [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52a37d1f-9e1d-0c11-8690-8a0590eb00dc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.630055] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1090.630391] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-553d2f73-c31b-440a-89e2-79cd2f4d2319 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.637770] env[62965]: DEBUG oslo_vmware.api [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 1090.637770] env[62965]: value = "task-1390395" [ 1090.637770] env[62965]: _type = "Task" [ 1090.637770] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.645295] env[62965]: DEBUG oslo_vmware.api [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390395, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.010009] env[62965]: DEBUG oslo_vmware.api [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52a37d1f-9e1d-0c11-8690-8a0590eb00dc, 'name': SearchDatastore_Task, 'duration_secs': 0.008715} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.010009] env[62965]: DEBUG oslo_concurrency.lockutils [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1091.010009] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1091.010195] env[62965]: DEBUG oslo_concurrency.lockutils [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1091.010346] env[62965]: DEBUG oslo_concurrency.lockutils [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1091.010529] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1091.010797] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a8ebac51-7690-4c4a-a068-e633ea0d8672 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.018444] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1091.018639] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1091.019307] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3417b22e-f521-4158-a977-0eb4382eb328 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.023995] env[62965]: DEBUG oslo_vmware.api [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Waiting for the task: (returnval){ [ 1091.023995] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52daaa28-c5d4-5f2a-e7c5-5e04a2051bc0" [ 1091.023995] env[62965]: _type = "Task" [ 1091.023995] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.031573] env[62965]: DEBUG oslo_vmware.api [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52daaa28-c5d4-5f2a-e7c5-5e04a2051bc0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.147814] env[62965]: DEBUG oslo_vmware.api [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390395, 'name': PowerOffVM_Task, 'duration_secs': 0.17765} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.148179] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1091.148419] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Updating instance 'b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7' progress to 17 {{(pid=62965) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1091.534028] env[62965]: DEBUG oslo_vmware.api [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52daaa28-c5d4-5f2a-e7c5-5e04a2051bc0, 'name': SearchDatastore_Task, 'duration_secs': 0.013694} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.534849] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7be4a1ac-b494-4d6c-b3ba-171d8c50424f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.539629] env[62965]: DEBUG oslo_vmware.api [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Waiting for the task: (returnval){ [ 1091.539629] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52734c4d-bfc7-864b-41d9-0371de49ab9d" [ 1091.539629] env[62965]: _type = "Task" [ 1091.539629] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.546958] env[62965]: DEBUG oslo_vmware.api [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52734c4d-bfc7-864b-41d9-0371de49ab9d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.655433] env[62965]: DEBUG nova.virt.hardware [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:24Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1091.655715] env[62965]: DEBUG nova.virt.hardware [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1091.655843] env[62965]: DEBUG nova.virt.hardware [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1091.656120] env[62965]: DEBUG nova.virt.hardware [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1091.656201] env[62965]: DEBUG nova.virt.hardware [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1091.656339] env[62965]: DEBUG nova.virt.hardware [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1091.656548] env[62965]: DEBUG nova.virt.hardware [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1091.656714] env[62965]: DEBUG nova.virt.hardware [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1091.656884] env[62965]: DEBUG nova.virt.hardware [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1091.657065] env[62965]: DEBUG nova.virt.hardware [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1091.657249] env[62965]: DEBUG nova.virt.hardware [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1091.662342] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e7ca80ae-5fcb-4df0-8876-866253ede0d2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.678648] env[62965]: DEBUG oslo_vmware.api [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 1091.678648] env[62965]: value = "task-1390396" [ 1091.678648] env[62965]: _type = "Task" [ 1091.678648] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.686555] env[62965]: DEBUG oslo_vmware.api [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390396, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.050145] env[62965]: DEBUG oslo_vmware.api [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52734c4d-bfc7-864b-41d9-0371de49ab9d, 'name': SearchDatastore_Task, 'duration_secs': 0.009669} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.050532] env[62965]: DEBUG oslo_concurrency.lockutils [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1092.050880] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 182e796f-2fbf-4304-9339-dc6e9b6b5839/182e796f-2fbf-4304-9339-dc6e9b6b5839.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1092.051224] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e33cc610-90ac-4f73-af1b-0eafa4269f04 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.059058] env[62965]: DEBUG oslo_vmware.api [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Waiting for the task: (returnval){ [ 1092.059058] env[62965]: value = "task-1390397" [ 1092.059058] env[62965]: _type = "Task" [ 1092.059058] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.067224] env[62965]: DEBUG oslo_vmware.api [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': task-1390397, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.187715] env[62965]: DEBUG oslo_vmware.api [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390396, 'name': ReconfigVM_Task, 'duration_secs': 0.155536} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.188050] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Updating instance 'b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7' progress to 33 {{(pid=62965) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1092.568997] env[62965]: DEBUG oslo_vmware.api [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': task-1390397, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.502909} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.569468] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] 182e796f-2fbf-4304-9339-dc6e9b6b5839/182e796f-2fbf-4304-9339-dc6e9b6b5839.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1092.569570] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1092.569756] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-06cd8847-026c-4b5b-ad2b-9cee666ed90c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.575709] env[62965]: DEBUG oslo_vmware.api [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Waiting for the task: (returnval){ [ 1092.575709] env[62965]: value = "task-1390398" [ 1092.575709] env[62965]: _type = "Task" [ 1092.575709] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.584094] env[62965]: DEBUG oslo_vmware.api [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': task-1390398, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.695723] env[62965]: DEBUG nova.virt.hardware [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1092.696136] env[62965]: DEBUG nova.virt.hardware [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1092.696468] env[62965]: DEBUG nova.virt.hardware [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1092.696832] env[62965]: DEBUG nova.virt.hardware [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1092.697160] env[62965]: DEBUG nova.virt.hardware [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1092.697492] env[62965]: DEBUG nova.virt.hardware [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1092.697883] env[62965]: DEBUG nova.virt.hardware [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1092.698197] env[62965]: DEBUG nova.virt.hardware [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1092.698528] env[62965]: DEBUG nova.virt.hardware [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1092.698876] env[62965]: DEBUG nova.virt.hardware [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1092.699187] env[62965]: DEBUG nova.virt.hardware [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1092.708770] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Reconfiguring VM instance instance-00000069 to detach disk 2000 {{(pid=62965) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1092.709208] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5b6ffa05-acc7-4fa2-8b32-d81e60d09195 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.733050] env[62965]: DEBUG oslo_vmware.api [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 1092.733050] env[62965]: value = "task-1390399" [ 1092.733050] env[62965]: _type = "Task" [ 1092.733050] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.741018] env[62965]: DEBUG oslo_vmware.api [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390399, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.085700] env[62965]: DEBUG oslo_vmware.api [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': task-1390398, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.210983} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.085976] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1093.086801] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2b33948-4845-445d-8973-df8ba54555a8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.106378] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] 182e796f-2fbf-4304-9339-dc6e9b6b5839/182e796f-2fbf-4304-9339-dc6e9b6b5839.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1093.106642] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec3e7e81-413e-4ac5-9903-007a9af47536 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.124973] env[62965]: DEBUG oslo_vmware.api [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Waiting for the task: (returnval){ [ 1093.124973] env[62965]: value = "task-1390400" [ 1093.124973] env[62965]: _type = "Task" [ 1093.124973] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.132405] env[62965]: DEBUG oslo_vmware.api [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': task-1390400, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.242702] env[62965]: DEBUG oslo_vmware.api [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390399, 'name': ReconfigVM_Task, 'duration_secs': 0.174114} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.243024] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Reconfigured VM instance instance-00000069 to detach disk 2000 {{(pid=62965) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1093.243799] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-963b2532-e6cd-4822-a54b-4ece12337f67 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.266147] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7/b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7.vmdk or device None with type thin {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1093.266461] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6e3aabca-48f9-4798-b228-bfac6672a206 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.284832] env[62965]: DEBUG oslo_vmware.api [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 1093.284832] env[62965]: value = "task-1390401" [ 1093.284832] env[62965]: _type = "Task" [ 1093.284832] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.292342] env[62965]: DEBUG oslo_vmware.api [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390401, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.634802] env[62965]: DEBUG oslo_vmware.api [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': task-1390400, 'name': ReconfigVM_Task, 'duration_secs': 0.273948} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.635114] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Reconfigured VM instance instance-0000006b to attach disk [datastore2] 182e796f-2fbf-4304-9339-dc6e9b6b5839/182e796f-2fbf-4304-9339-dc6e9b6b5839.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1093.635727] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9b9eeba4-9842-4cc9-a188-4f69675e0d4a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.641790] env[62965]: DEBUG oslo_vmware.api [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Waiting for the task: (returnval){ [ 1093.641790] env[62965]: value = "task-1390402" [ 1093.641790] env[62965]: _type = "Task" [ 1093.641790] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.650247] env[62965]: DEBUG oslo_vmware.api [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': task-1390402, 'name': Rename_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.794121] env[62965]: DEBUG oslo_vmware.api [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390401, 'name': ReconfigVM_Task, 'duration_secs': 0.348054} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.794416] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Reconfigured VM instance instance-00000069 to attach disk [datastore1] b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7/b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7.vmdk or device None with type thin {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1093.794690] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Updating instance 'b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7' progress to 50 {{(pid=62965) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1094.104948] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1094.104948] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1094.151474] env[62965]: DEBUG oslo_vmware.api [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': task-1390402, 'name': Rename_Task, 'duration_secs': 0.133288} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.151960] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1094.152208] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a8321b1d-374c-43fc-8174-5db35faa58af {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.158435] env[62965]: DEBUG oslo_vmware.api [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Waiting for the task: (returnval){ [ 1094.158435] env[62965]: value = "task-1390403" [ 1094.158435] env[62965]: _type = "Task" [ 1094.158435] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.165721] env[62965]: DEBUG oslo_vmware.api [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': task-1390403, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.234753] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquiring lock "73a7ca0e-b1d3-4464-969b-5111d2506a8b" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1094.235025] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "73a7ca0e-b1d3-4464-969b-5111d2506a8b" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1094.235280] env[62965]: INFO nova.compute.manager [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Shelving [ 1094.302248] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d3a3345-9b5d-49a5-80ec-9891d52de67a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.321861] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e82d00c-eb12-4306-89e8-7d529e2cdc94 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.342637] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Updating instance 'b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7' progress to 67 {{(pid=62965) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1094.610501] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1094.610739] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Starting heal instance info cache {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10257}} [ 1094.668098] env[62965]: DEBUG oslo_vmware.api [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': task-1390403, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.887888] env[62965]: DEBUG nova.network.neutron [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Port 5232b21c-f849-45de-916a-788114415557 binding to destination host cpu-1 is already ACTIVE {{(pid=62965) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1095.172199] env[62965]: DEBUG oslo_vmware.api [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': task-1390403, 'name': PowerOnVM_Task, 'duration_secs': 0.979696} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.172199] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1095.172199] env[62965]: DEBUG nova.compute.manager [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1095.172199] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cdd80e9-3bec-43de-b169-14176c2e10f9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.244923] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1095.245269] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ad3b8b99-b515-4475-a084-9d6e9903be5f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.253716] env[62965]: DEBUG oslo_vmware.api [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 1095.253716] env[62965]: value = "task-1390404" [ 1095.253716] env[62965]: _type = "Task" [ 1095.253716] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.261170] env[62965]: DEBUG oslo_vmware.api [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390404, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.691830] env[62965]: DEBUG oslo_concurrency.lockutils [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.692132] env[62965]: DEBUG oslo_concurrency.lockutils [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.692325] env[62965]: DEBUG nova.objects.instance [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62965) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1095.763551] env[62965]: DEBUG oslo_vmware.api [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390404, 'name': PowerOffVM_Task, 'duration_secs': 0.199411} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.764487] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1095.764850] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0771d68-1003-4c9b-87bd-d2d19159847e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.783434] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c489bba-a4f1-4a3b-a39a-2412f63c1ee6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.912480] env[62965]: DEBUG oslo_concurrency.lockutils [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.912743] env[62965]: DEBUG oslo_concurrency.lockutils [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.912951] env[62965]: DEBUG oslo_concurrency.lockutils [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.978762] env[62965]: DEBUG oslo_concurrency.lockutils [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Acquiring lock "182e796f-2fbf-4304-9339-dc6e9b6b5839" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.978963] env[62965]: DEBUG oslo_concurrency.lockutils [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Lock "182e796f-2fbf-4304-9339-dc6e9b6b5839" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.979190] env[62965]: DEBUG oslo_concurrency.lockutils [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Acquiring lock "182e796f-2fbf-4304-9339-dc6e9b6b5839-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.979414] env[62965]: DEBUG oslo_concurrency.lockutils [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Lock "182e796f-2fbf-4304-9339-dc6e9b6b5839-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.979551] env[62965]: DEBUG oslo_concurrency.lockutils [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Lock "182e796f-2fbf-4304-9339-dc6e9b6b5839-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.981643] env[62965]: INFO nova.compute.manager [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Terminating instance [ 1096.294061] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Creating Snapshot of the VM instance {{(pid=62965) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1096.294342] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-f299cf72-e074-4341-9eae-02c57fb741ed {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.302259] env[62965]: DEBUG oslo_vmware.api [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 1096.302259] env[62965]: value = "task-1390405" [ 1096.302259] env[62965]: _type = "Task" [ 1096.302259] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.310159] env[62965]: DEBUG oslo_vmware.api [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390405, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.485394] env[62965]: DEBUG oslo_concurrency.lockutils [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Acquiring lock "refresh_cache-182e796f-2fbf-4304-9339-dc6e9b6b5839" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1096.485585] env[62965]: DEBUG oslo_concurrency.lockutils [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Acquired lock "refresh_cache-182e796f-2fbf-4304-9339-dc6e9b6b5839" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.485754] env[62965]: DEBUG nova.network.neutron [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1096.700071] env[62965]: DEBUG oslo_concurrency.lockutils [None req-74eaddc7-28bd-43bf-80a3-f259228973c1 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.008s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.811926] env[62965]: DEBUG oslo_vmware.api [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390405, 'name': CreateSnapshot_Task, 'duration_secs': 0.423329} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.812204] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Created Snapshot of the VM instance {{(pid=62965) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1096.812932] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b1dc151-6f94-47a4-9997-fff2763f4eef {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.967920] env[62965]: DEBUG oslo_concurrency.lockutils [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "refresh_cache-b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1096.968145] env[62965]: DEBUG oslo_concurrency.lockutils [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquired lock "refresh_cache-b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.968331] env[62965]: DEBUG nova.network.neutron [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1097.002487] env[62965]: DEBUG nova.network.neutron [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1097.054535] env[62965]: DEBUG nova.network.neutron [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1097.125775] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Didn't find any instances for network info cache update. {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10343}} [ 1097.125997] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.126166] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.126321] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.126470] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.126613] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.126759] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.126888] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62965) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10876}} [ 1097.127044] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1097.329655] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Creating linked-clone VM from snapshot {{(pid=62965) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1097.329993] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-3ccaf2bc-1d6e-4a62-a2be-9a1656fec987 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.339087] env[62965]: DEBUG oslo_vmware.api [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 1097.339087] env[62965]: value = "task-1390406" [ 1097.339087] env[62965]: _type = "Task" [ 1097.339087] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.347055] env[62965]: DEBUG oslo_vmware.api [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390406, 'name': CloneVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.556777] env[62965]: DEBUG oslo_concurrency.lockutils [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Releasing lock "refresh_cache-182e796f-2fbf-4304-9339-dc6e9b6b5839" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1097.557220] env[62965]: DEBUG nova.compute.manager [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1097.557423] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1097.558294] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-540d37cf-c606-4e86-b0d0-454dcb9ea2dc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.566745] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1097.567021] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e00a1495-5e87-49bc-aa98-74807dd6c4b5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.575069] env[62965]: DEBUG oslo_vmware.api [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Waiting for the task: (returnval){ [ 1097.575069] env[62965]: value = "task-1390407" [ 1097.575069] env[62965]: _type = "Task" [ 1097.575069] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.582857] env[62965]: DEBUG oslo_vmware.api [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': task-1390407, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.629635] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.630163] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.630287] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.630510] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62965) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1097.631873] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c617e7f-6bc2-4d5f-b697-e1201ad81e73 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.643088] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8c71c5e-5b84-47b5-83f8-754345dcef82 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.667130] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3af2b608-57f2-4852-ba14-75ca158dd771 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.676971] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31b2900c-3d92-4eca-a1e3-4124392c8052 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.724552] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180929MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62965) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1097.724854] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.725160] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.765214] env[62965]: DEBUG nova.network.neutron [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Updating instance_info_cache with network_info: [{"id": "5232b21c-f849-45de-916a-788114415557", "address": "fa:16:3e:51:9d:68", "network": {"id": "62052700-1781-4933-93c8-fa1aafdb0fd2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1931798228-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "978666785b0743bb843de96585f96fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5232b21c-f8", "ovs_interfaceid": "5232b21c-f849-45de-916a-788114415557", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1097.849502] env[62965]: DEBUG oslo_vmware.api [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390406, 'name': CloneVM_Task} progress is 94%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.083956] env[62965]: DEBUG oslo_vmware.api [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': task-1390407, 'name': PowerOffVM_Task, 'duration_secs': 0.173061} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.084277] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1098.084464] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1098.084731] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dd11d174-04fb-4b2a-804f-b11b9567b29a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.111130] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1098.111356] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Deleting contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1098.111584] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Deleting the datastore file [datastore2] 182e796f-2fbf-4304-9339-dc6e9b6b5839 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1098.111954] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-80c37668-9736-469e-a993-ee2befd58b9a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.118185] env[62965]: DEBUG oslo_vmware.api [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Waiting for the task: (returnval){ [ 1098.118185] env[62965]: value = "task-1390409" [ 1098.118185] env[62965]: _type = "Task" [ 1098.118185] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.125637] env[62965]: DEBUG oslo_vmware.api [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': task-1390409, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.268383] env[62965]: DEBUG oslo_concurrency.lockutils [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Releasing lock "refresh_cache-b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1098.348612] env[62965]: DEBUG oslo_vmware.api [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390406, 'name': CloneVM_Task, 'duration_secs': 0.917871} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.348878] env[62965]: INFO nova.virt.vmwareapi.vmops [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Created linked-clone VM from snapshot [ 1098.349616] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c1aeb2a-b5db-4131-acbf-f6c97878e1a6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.357084] env[62965]: DEBUG nova.virt.vmwareapi.images [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Uploading image f1265e9a-63c3-4300-9384-70985cea3f2f {{(pid=62965) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1098.417682] env[62965]: DEBUG oslo_vmware.rw_handles [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1098.417682] env[62965]: value = "vm-295065" [ 1098.417682] env[62965]: _type = "VirtualMachine" [ 1098.417682] env[62965]: }. {{(pid=62965) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1098.418951] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-59ce730a-c1c5-4c10-93e4-7fbbbd766753 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.426393] env[62965]: DEBUG oslo_vmware.rw_handles [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lease: (returnval){ [ 1098.426393] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52f8a5cd-7c39-b5c6-4bbb-699a42d494d2" [ 1098.426393] env[62965]: _type = "HttpNfcLease" [ 1098.426393] env[62965]: } obtained for exporting VM: (result){ [ 1098.426393] env[62965]: value = "vm-295065" [ 1098.426393] env[62965]: _type = "VirtualMachine" [ 1098.426393] env[62965]: }. {{(pid=62965) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1098.426393] env[62965]: DEBUG oslo_vmware.api [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the lease: (returnval){ [ 1098.426393] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52f8a5cd-7c39-b5c6-4bbb-699a42d494d2" [ 1098.426393] env[62965]: _type = "HttpNfcLease" [ 1098.426393] env[62965]: } to be ready. {{(pid=62965) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1098.432544] env[62965]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1098.432544] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52f8a5cd-7c39-b5c6-4bbb-699a42d494d2" [ 1098.432544] env[62965]: _type = "HttpNfcLease" [ 1098.432544] env[62965]: } is initializing. {{(pid=62965) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1098.627755] env[62965]: DEBUG oslo_vmware.api [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Task: {'id': task-1390409, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.087704} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.628121] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1098.628322] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Deleted contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1098.628506] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1098.628680] env[62965]: INFO nova.compute.manager [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Took 1.07 seconds to destroy the instance on the hypervisor. [ 1098.628923] env[62965]: DEBUG oslo.service.loopingcall [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1098.629140] env[62965]: DEBUG nova.compute.manager [-] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1098.629238] env[62965]: DEBUG nova.network.neutron [-] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1098.643646] env[62965]: DEBUG nova.network.neutron [-] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1098.736340] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Applying migration context for instance b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7 as it has an incoming, in-progress migration 330e662e-d3f2-4c7d-84ec-bf24984e877a. Migration status is post-migrating {{(pid=62965) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1098.737266] env[62965]: INFO nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Updating resource usage from migration 330e662e-d3f2-4c7d-84ec-bf24984e877a [ 1098.933286] env[62965]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1098.933286] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52f8a5cd-7c39-b5c6-4bbb-699a42d494d2" [ 1098.933286] env[62965]: _type = "HttpNfcLease" [ 1098.933286] env[62965]: } is ready. {{(pid=62965) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1098.934790] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 924c5cdc-0450-4ced-b920-50e2d5060fd2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1098.934928] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 73a7ca0e-b1d3-4464-969b-5111d2506a8b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1098.935064] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 385648d7-20e3-442c-b6a1-3d972d266487 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1098.935186] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 182e796f-2fbf-4304-9339-dc6e9b6b5839 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1098.935302] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Migration 330e662e-d3f2-4c7d-84ec-bf24984e877a is active on this compute host and has allocations in placement: {'resources': {'VCPU': 1, 'MEMORY_MB': 192, 'DISK_GB': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1098.935416] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1098.935598] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=62965) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1098.935733] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1728MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=62965) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1098.937889] env[62965]: DEBUG oslo_vmware.rw_handles [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1098.937889] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52f8a5cd-7c39-b5c6-4bbb-699a42d494d2" [ 1098.937889] env[62965]: _type = "HttpNfcLease" [ 1098.937889] env[62965]: }. {{(pid=62965) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1098.938622] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd87e9d6-9a8f-4ade-8194-b56aac6633df {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.941712] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed324e02-8bdc-4d52-8112-e23554549ed8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.949139] env[62965]: DEBUG oslo_vmware.rw_handles [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a59cc7-a7f0-bc68-621e-0a0b98e49580/disk-0.vmdk from lease info. {{(pid=62965) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1098.949325] env[62965]: DEBUG oslo_vmware.rw_handles [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a59cc7-a7f0-bc68-621e-0a0b98e49580/disk-0.vmdk for reading. {{(pid=62965) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1098.967014] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4c71f4f-c01c-47f7-b88e-043da6b60b00 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.030105] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Updating instance 'b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7' progress to 83 {{(pid=62965) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1099.060284] env[62965]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-1ceac076-0e5b-49b3-808f-7e3234f7a0bf {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.077498] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69e9af95-da0d-49cf-90ce-ff3331ba9463 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.091856] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6484cf2d-a1ea-4fd1-947f-eed4155217ef {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.124011] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-211d980b-0900-4607-941e-b938cd6ac87c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.131463] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e83bc52e-e63c-4626-b577-5e549c6ecc9b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.144807] env[62965]: DEBUG nova.compute.provider_tree [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1099.146007] env[62965]: DEBUG nova.network.neutron [-] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1099.536753] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1099.537143] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-741bf096-d744-4695-aa37-645e89fee77f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.544876] env[62965]: DEBUG oslo_vmware.api [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 1099.544876] env[62965]: value = "task-1390411" [ 1099.544876] env[62965]: _type = "Task" [ 1099.544876] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.553353] env[62965]: DEBUG oslo_vmware.api [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390411, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.648749] env[62965]: DEBUG nova.scheduler.client.report [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1099.652383] env[62965]: INFO nova.compute.manager [-] [instance: 182e796f-2fbf-4304-9339-dc6e9b6b5839] Took 1.02 seconds to deallocate network for instance. [ 1100.058223] env[62965]: DEBUG oslo_vmware.api [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390411, 'name': PowerOnVM_Task, 'duration_secs': 0.404192} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.058699] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1100.059092] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-286963b6-b05f-4b23-9bf9-60e155873ace tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Updating instance 'b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7' progress to 100 {{(pid=62965) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1100.154939] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62965) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1100.155171] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.430s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.155475] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1100.155717] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Cleaning up deleted instances {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11557}} [ 1100.160857] env[62965]: DEBUG oslo_concurrency.lockutils [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.161166] env[62965]: DEBUG oslo_concurrency.lockutils [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1100.161656] env[62965]: DEBUG nova.objects.instance [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Lazy-loading 'resources' on Instance uuid 182e796f-2fbf-4304-9339-dc6e9b6b5839 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1100.668386] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] There are 40 instances to clean {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11566}} [ 1100.668764] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 740b4d83-afc5-4ee2-9130-8c1d0697e5e5] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1100.785735] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77dfe33c-313a-4231-a5e3-052d6ea75915 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.794641] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12169f62-6ad2-4f43-9334-8e671b341735 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.825540] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aec8ee8-0af3-4870-a86e-28c773bedfe3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.833888] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1f7aedc-a1a8-462a-b81b-cc8a76161248 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.848167] env[62965]: DEBUG nova.compute.provider_tree [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1101.174295] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 88e45c29-4d21-4f91-92f4-fdebaf905ccb] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1101.351866] env[62965]: DEBUG nova.scheduler.client.report [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1101.677501] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 58c62ac1-203c-4370-8c94-420d655b7c08] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1101.815676] env[62965]: DEBUG oslo_concurrency.lockutils [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.815939] env[62965]: DEBUG oslo_concurrency.lockutils [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.816147] env[62965]: DEBUG nova.compute.manager [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Going to confirm migration 4 {{(pid=62965) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5112}} [ 1101.856445] env[62965]: DEBUG oslo_concurrency.lockutils [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.695s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.885786] env[62965]: INFO nova.scheduler.client.report [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Deleted allocations for instance 182e796f-2fbf-4304-9339-dc6e9b6b5839 [ 1102.181061] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: df03e327-959c-42af-85cd-a7af2ec4d0f7] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1102.374973] env[62965]: DEBUG oslo_concurrency.lockutils [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "refresh_cache-b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1102.375185] env[62965]: DEBUG oslo_concurrency.lockutils [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquired lock "refresh_cache-b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1102.375368] env[62965]: DEBUG nova.network.neutron [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1102.375642] env[62965]: DEBUG nova.objects.instance [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lazy-loading 'info_cache' on Instance uuid b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1102.393177] env[62965]: DEBUG oslo_concurrency.lockutils [None req-354292ff-30df-475d-8e0a-2aa539a229b9 tempest-ServerShowV254Test-519200351 tempest-ServerShowV254Test-519200351-project-member] Lock "182e796f-2fbf-4304-9339-dc6e9b6b5839" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.414s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.685150] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 913c0536-5e39-4654-b083-2703a6aed328] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1103.189180] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 9941c75d-3f4a-4fb3-b3dd-a0b59de7d11b] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1103.607459] env[62965]: DEBUG nova.network.neutron [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Updating instance_info_cache with network_info: [{"id": "5232b21c-f849-45de-916a-788114415557", "address": "fa:16:3e:51:9d:68", "network": {"id": "62052700-1781-4933-93c8-fa1aafdb0fd2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1931798228-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "978666785b0743bb843de96585f96fb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5232b21c-f8", "ovs_interfaceid": "5232b21c-f849-45de-916a-788114415557", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1103.692949] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 8ae1fde5-e94e-43ed-816a-9e22633dfdd1] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1104.110554] env[62965]: DEBUG oslo_concurrency.lockutils [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Releasing lock "refresh_cache-b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1104.110813] env[62965]: DEBUG nova.objects.instance [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lazy-loading 'migration_context' on Instance uuid b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1104.196688] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 0620c47c-2918-43dc-9e67-f32ca3dd20f1] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1104.614670] env[62965]: DEBUG nova.objects.base [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62965) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1104.615702] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f34de406-ee1d-4ac7-8f20-5ff291b67768 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.636812] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b85393c4-9eeb-4ad6-8b9d-095103035abf {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.643501] env[62965]: DEBUG oslo_vmware.api [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 1104.643501] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]523b93fc-9ff5-52b1-fbe6-ef55c415d407" [ 1104.643501] env[62965]: _type = "Task" [ 1104.643501] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.652201] env[62965]: DEBUG oslo_vmware.api [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]523b93fc-9ff5-52b1-fbe6-ef55c415d407, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.700021] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 8efad776-6cbe-49ab-83dc-a410d8417505] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1105.157084] env[62965]: DEBUG oslo_vmware.api [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]523b93fc-9ff5-52b1-fbe6-ef55c415d407, 'name': SearchDatastore_Task, 'duration_secs': 0.010796} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.157447] env[62965]: DEBUG oslo_concurrency.lockutils [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.157779] env[62965]: DEBUG oslo_concurrency.lockutils [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.203656] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: fe0559f0-d1ac-4af8-b554-2d31c38c2e73] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1105.707329] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 43db5f53-a719-4384-a8d2-63721dde0187] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1105.738830] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-726a3e66-e474-4948-b9fe-e536579c7fb5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.747417] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1745b675-298f-46df-b6b7-9c0423a22066 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.779067] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f447cd14-db1d-427d-9141-440bc12e3667 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.787818] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49a16f2b-cf20-4e16-b1aa-6763895f975b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.804432] env[62965]: DEBUG nova.compute.provider_tree [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1106.210590] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 8a54b436-3b43-439e-bf6d-417db1fa66b2] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1106.308046] env[62965]: DEBUG nova.scheduler.client.report [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1106.714247] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 464a3a31-06bf-4662-a5b1-d185d9a64f0b] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1107.217754] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 08f7b05f-3fb5-48ac-a68a-1d6fd804a622] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1107.318922] env[62965]: DEBUG oslo_concurrency.lockutils [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.161s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.720666] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: a63b301d-8d52-4fec-951f-a15cda9449ac] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1107.769222] env[62965]: DEBUG oslo_vmware.rw_handles [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a59cc7-a7f0-bc68-621e-0a0b98e49580/disk-0.vmdk. {{(pid=62965) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1107.770153] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72908876-d73c-4cd8-9ae5-fca0f5a01bd8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.777266] env[62965]: DEBUG oslo_vmware.rw_handles [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a59cc7-a7f0-bc68-621e-0a0b98e49580/disk-0.vmdk is in state: ready. {{(pid=62965) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1107.777429] env[62965]: ERROR oslo_vmware.rw_handles [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a59cc7-a7f0-bc68-621e-0a0b98e49580/disk-0.vmdk due to incomplete transfer. [ 1107.777647] env[62965]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-ebd26c9e-975a-4b63-a068-95a7ff78af9a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.785606] env[62965]: DEBUG oslo_vmware.rw_handles [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52a59cc7-a7f0-bc68-621e-0a0b98e49580/disk-0.vmdk. {{(pid=62965) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1107.785801] env[62965]: DEBUG nova.virt.vmwareapi.images [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Uploaded image f1265e9a-63c3-4300-9384-70985cea3f2f to the Glance image server {{(pid=62965) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1107.788074] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Destroying the VM {{(pid=62965) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1107.788314] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-09ac73ba-ca8c-4aef-b5c7-c44486d44c96 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.794775] env[62965]: DEBUG oslo_vmware.api [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 1107.794775] env[62965]: value = "task-1390412" [ 1107.794775] env[62965]: _type = "Task" [ 1107.794775] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.803333] env[62965]: DEBUG oslo_vmware.api [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390412, 'name': Destroy_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.877903] env[62965]: INFO nova.scheduler.client.report [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Deleted allocation for migration 330e662e-d3f2-4c7d-84ec-bf24984e877a [ 1108.224064] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 3223b1ed-abc1-439c-bcad-6317eb204e45] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1108.304845] env[62965]: DEBUG oslo_vmware.api [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390412, 'name': Destroy_Task, 'duration_secs': 0.349149} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.305166] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Destroyed the VM [ 1108.305419] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Deleting Snapshot of the VM instance {{(pid=62965) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1108.305685] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-9815a978-faf3-4465-86f9-d63766b5c632 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.312094] env[62965]: DEBUG oslo_vmware.api [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 1108.312094] env[62965]: value = "task-1390413" [ 1108.312094] env[62965]: _type = "Task" [ 1108.312094] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.320204] env[62965]: DEBUG oslo_vmware.api [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390413, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.384030] env[62965]: DEBUG oslo_concurrency.lockutils [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.568s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.727693] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 97954de1-7b4b-4615-8d0f-0e8580a515ed] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1108.822359] env[62965]: DEBUG oslo_vmware.api [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390413, 'name': RemoveSnapshot_Task, 'duration_secs': 0.325762} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.822627] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Deleted Snapshot of the VM instance {{(pid=62965) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1108.822908] env[62965]: DEBUG nova.compute.manager [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1108.823676] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edb9105f-c191-439d-a0ce-ce34c454d478 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.231288] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 1be808db-d619-4c56-8092-ca52809ec10e] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1109.337493] env[62965]: INFO nova.compute.manager [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Shelve offloading [ 1109.735381] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 0eb943ae-462c-40dd-a870-7a884ea2db74] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1109.839756] env[62965]: DEBUG oslo_concurrency.lockutils [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.840036] env[62965]: DEBUG oslo_concurrency.lockutils [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1109.840297] env[62965]: DEBUG oslo_concurrency.lockutils [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.840501] env[62965]: DEBUG oslo_concurrency.lockutils [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1109.840676] env[62965]: DEBUG oslo_concurrency.lockutils [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.842473] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1109.842745] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-96e438a6-4e24-45fa-a7b8-2ab23ee21368 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.845107] env[62965]: INFO nova.compute.manager [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Terminating instance [ 1109.854139] env[62965]: DEBUG oslo_vmware.api [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 1109.854139] env[62965]: value = "task-1390414" [ 1109.854139] env[62965]: _type = "Task" [ 1109.854139] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.864090] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] VM already powered off {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1109.864297] env[62965]: DEBUG nova.compute.manager [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1109.865071] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bf20e93-711b-4e7c-b199-884f3e07337e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.870610] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquiring lock "refresh_cache-73a7ca0e-b1d3-4464-969b-5111d2506a8b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1109.870777] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquired lock "refresh_cache-73a7ca0e-b1d3-4464-969b-5111d2506a8b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.870946] env[62965]: DEBUG nova.network.neutron [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1110.239091] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 00f24976-2f0e-4a2f-8f2e-012b0fce479f] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1110.349097] env[62965]: DEBUG nova.compute.manager [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1110.349344] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1110.350221] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-466afaa1-1fbf-4123-8339-cb8d2cdc439f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.358145] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1110.358395] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-db59b4e0-c630-438f-ac37-c08e12933654 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.364687] env[62965]: DEBUG oslo_vmware.api [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 1110.364687] env[62965]: value = "task-1390415" [ 1110.364687] env[62965]: _type = "Task" [ 1110.364687] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.375167] env[62965]: DEBUG oslo_vmware.api [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390415, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.678503] env[62965]: DEBUG nova.network.neutron [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Updating instance_info_cache with network_info: [{"id": "5f939a36-bc5f-4fd7-b29e-011f72b3256c", "address": "fa:16:3e:e3:06:99", "network": {"id": "49dd42b0-6397-4fdd-a319-ed920c72a5af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-373172566-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d91c99b14018469e8998fd1ccc0fab4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f939a36-bc", "ovs_interfaceid": "5f939a36-bc5f-4fd7-b29e-011f72b3256c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.741795] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: fbbcbe60-92f3-497d-b190-520a2547b3a9] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1110.874367] env[62965]: DEBUG oslo_vmware.api [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390415, 'name': PowerOffVM_Task, 'duration_secs': 0.209341} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.874630] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1110.874801] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1110.875064] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c2cb74c0-e370-499c-bfb2-d9a26f2d33e9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.938648] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1110.938864] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Deleting contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1110.939063] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Deleting the datastore file [datastore1] b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1110.939339] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3de67c44-ffe0-4599-9f9d-e0a8e1bc26fe {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.946412] env[62965]: DEBUG oslo_vmware.api [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for the task: (returnval){ [ 1110.946412] env[62965]: value = "task-1390417" [ 1110.946412] env[62965]: _type = "Task" [ 1110.946412] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.955264] env[62965]: DEBUG oslo_vmware.api [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390417, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.181794] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Releasing lock "refresh_cache-73a7ca0e-b1d3-4464-969b-5111d2506a8b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.244636] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: e0c4e188-a041-4aa7-95de-6e8718b17ade] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1111.387998] env[62965]: DEBUG nova.compute.manager [req-b3b571cc-c249-4d37-8d64-5ccb0917af12 req-f63e1e55-994b-4e6b-9f04-5feb255cc901 service nova] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Received event network-vif-unplugged-5f939a36-bc5f-4fd7-b29e-011f72b3256c {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1111.388284] env[62965]: DEBUG oslo_concurrency.lockutils [req-b3b571cc-c249-4d37-8d64-5ccb0917af12 req-f63e1e55-994b-4e6b-9f04-5feb255cc901 service nova] Acquiring lock "73a7ca0e-b1d3-4464-969b-5111d2506a8b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.388563] env[62965]: DEBUG oslo_concurrency.lockutils [req-b3b571cc-c249-4d37-8d64-5ccb0917af12 req-f63e1e55-994b-4e6b-9f04-5feb255cc901 service nova] Lock "73a7ca0e-b1d3-4464-969b-5111d2506a8b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.388738] env[62965]: DEBUG oslo_concurrency.lockutils [req-b3b571cc-c249-4d37-8d64-5ccb0917af12 req-f63e1e55-994b-4e6b-9f04-5feb255cc901 service nova] Lock "73a7ca0e-b1d3-4464-969b-5111d2506a8b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.388913] env[62965]: DEBUG nova.compute.manager [req-b3b571cc-c249-4d37-8d64-5ccb0917af12 req-f63e1e55-994b-4e6b-9f04-5feb255cc901 service nova] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] No waiting events found dispatching network-vif-unplugged-5f939a36-bc5f-4fd7-b29e-011f72b3256c {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1111.389112] env[62965]: WARNING nova.compute.manager [req-b3b571cc-c249-4d37-8d64-5ccb0917af12 req-f63e1e55-994b-4e6b-9f04-5feb255cc901 service nova] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Received unexpected event network-vif-unplugged-5f939a36-bc5f-4fd7-b29e-011f72b3256c for instance with vm_state shelved and task_state shelving_offloading. [ 1111.456639] env[62965]: DEBUG oslo_vmware.api [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Task: {'id': task-1390417, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140737} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.456860] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1111.457059] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Deleted contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1111.457244] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1111.457417] env[62965]: INFO nova.compute.manager [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1111.457657] env[62965]: DEBUG oslo.service.loopingcall [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1111.457848] env[62965]: DEBUG nova.compute.manager [-] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1111.457944] env[62965]: DEBUG nova.network.neutron [-] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1111.475063] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1111.475883] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eceaa65e-a233-4d71-9bad-413473133447 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.482919] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1111.483162] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-25bbb28f-aca9-48b8-841d-0fbc2f50ee51 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.561544] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1111.561761] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Deleting contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1111.561938] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Deleting the datastore file [datastore2] 73a7ca0e-b1d3-4464-969b-5111d2506a8b {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1111.562259] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d7871674-7479-4516-9228-c0839d452457 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.568308] env[62965]: DEBUG oslo_vmware.api [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 1111.568308] env[62965]: value = "task-1390419" [ 1111.568308] env[62965]: _type = "Task" [ 1111.568308] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.576105] env[62965]: DEBUG oslo_vmware.api [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390419, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.734203] env[62965]: DEBUG nova.compute.manager [req-a56776db-5375-4f3b-a7d4-74de48ac8aa4 req-64edc988-a70f-467b-80d5-ee3eb6edfdef service nova] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Received event network-vif-deleted-5232b21c-f849-45de-916a-788114415557 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1111.734346] env[62965]: INFO nova.compute.manager [req-a56776db-5375-4f3b-a7d4-74de48ac8aa4 req-64edc988-a70f-467b-80d5-ee3eb6edfdef service nova] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Neutron deleted interface 5232b21c-f849-45de-916a-788114415557; detaching it from the instance and deleting it from the info cache [ 1111.734574] env[62965]: DEBUG nova.network.neutron [req-a56776db-5375-4f3b-a7d4-74de48ac8aa4 req-64edc988-a70f-467b-80d5-ee3eb6edfdef service nova] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1111.747183] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: eed55d8e-21dc-47d4-a689-415ef6a3891d] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1112.078080] env[62965]: DEBUG oslo_vmware.api [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390419, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13573} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.078354] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1112.078541] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Deleted contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1112.078718] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1112.098520] env[62965]: INFO nova.scheduler.client.report [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Deleted allocations for instance 73a7ca0e-b1d3-4464-969b-5111d2506a8b [ 1112.214353] env[62965]: DEBUG nova.network.neutron [-] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1112.236623] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2b435840-94af-4c7d-af4e-24d9ff8a09fd {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.246173] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6dc308d-0b3f-44f3-9438-e8fa66f043ad {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.256113] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 892b1302-f536-4dcd-9227-a545ae1ac731] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1112.271018] env[62965]: DEBUG nova.compute.manager [req-a56776db-5375-4f3b-a7d4-74de48ac8aa4 req-64edc988-a70f-467b-80d5-ee3eb6edfdef service nova] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Detach interface failed, port_id=5232b21c-f849-45de-916a-788114415557, reason: Instance b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7 could not be found. {{(pid=62965) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11294}} [ 1112.603401] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1112.603687] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1112.603920] env[62965]: DEBUG nova.objects.instance [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lazy-loading 'resources' on Instance uuid 73a7ca0e-b1d3-4464-969b-5111d2506a8b {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1112.717209] env[62965]: INFO nova.compute.manager [-] [instance: b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7] Took 1.26 seconds to deallocate network for instance. [ 1112.759359] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: f82d86ec-3364-449d-be54-35627082c02b] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1113.108433] env[62965]: DEBUG nova.objects.instance [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lazy-loading 'numa_topology' on Instance uuid 73a7ca0e-b1d3-4464-969b-5111d2506a8b {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1113.224464] env[62965]: DEBUG oslo_concurrency.lockutils [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1113.263493] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: fa511b61-887e-46a1-aaf6-586d716c0fb5] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1113.538695] env[62965]: DEBUG nova.compute.manager [req-03052458-5ecf-41c0-8fcd-f9472d19f171 req-25c33daf-e106-48af-9810-b5e3fd6ab141 service nova] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Received event network-changed-5f939a36-bc5f-4fd7-b29e-011f72b3256c {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1113.538953] env[62965]: DEBUG nova.compute.manager [req-03052458-5ecf-41c0-8fcd-f9472d19f171 req-25c33daf-e106-48af-9810-b5e3fd6ab141 service nova] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Refreshing instance network info cache due to event network-changed-5f939a36-bc5f-4fd7-b29e-011f72b3256c. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 1113.539334] env[62965]: DEBUG oslo_concurrency.lockutils [req-03052458-5ecf-41c0-8fcd-f9472d19f171 req-25c33daf-e106-48af-9810-b5e3fd6ab141 service nova] Acquiring lock "refresh_cache-73a7ca0e-b1d3-4464-969b-5111d2506a8b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1113.539334] env[62965]: DEBUG oslo_concurrency.lockutils [req-03052458-5ecf-41c0-8fcd-f9472d19f171 req-25c33daf-e106-48af-9810-b5e3fd6ab141 service nova] Acquired lock "refresh_cache-73a7ca0e-b1d3-4464-969b-5111d2506a8b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1113.539476] env[62965]: DEBUG nova.network.neutron [req-03052458-5ecf-41c0-8fcd-f9472d19f171 req-25c33daf-e106-48af-9810-b5e3fd6ab141 service nova] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Refreshing network info cache for port 5f939a36-bc5f-4fd7-b29e-011f72b3256c {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1113.610842] env[62965]: DEBUG nova.objects.base [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Object Instance<73a7ca0e-b1d3-4464-969b-5111d2506a8b> lazy-loaded attributes: resources,numa_topology {{(pid=62965) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1113.664583] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e59c8db-4e64-43a2-99d0-1238f63dd724 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.672897] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a581989-d72c-43b9-9a3c-f1a26fb0a205 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.703589] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09de45c2-7e1f-4e6c-9654-98fe0db0ba52 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.711031] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69344055-08a2-47ce-abe1-1b3396b8544b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.723714] env[62965]: DEBUG nova.compute.provider_tree [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1113.766506] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: d038608c-9264-4fd5-b9ba-0653c98ec0e4] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1114.016205] env[62965]: DEBUG oslo_concurrency.lockutils [None req-06f31f85-a32b-46ae-bfb9-427281418d85 tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Acquiring lock "924c5cdc-0450-4ced-b920-50e2d5060fd2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.016562] env[62965]: DEBUG oslo_concurrency.lockutils [None req-06f31f85-a32b-46ae-bfb9-427281418d85 tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Lock "924c5cdc-0450-4ced-b920-50e2d5060fd2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.016778] env[62965]: DEBUG oslo_concurrency.lockutils [None req-06f31f85-a32b-46ae-bfb9-427281418d85 tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Acquiring lock "924c5cdc-0450-4ced-b920-50e2d5060fd2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.016965] env[62965]: DEBUG oslo_concurrency.lockutils [None req-06f31f85-a32b-46ae-bfb9-427281418d85 tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Lock "924c5cdc-0450-4ced-b920-50e2d5060fd2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.019015] env[62965]: DEBUG oslo_concurrency.lockutils [None req-06f31f85-a32b-46ae-bfb9-427281418d85 tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Lock "924c5cdc-0450-4ced-b920-50e2d5060fd2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.019479] env[62965]: INFO nova.compute.manager [None req-06f31f85-a32b-46ae-bfb9-427281418d85 tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Terminating instance [ 1114.226742] env[62965]: DEBUG nova.scheduler.client.report [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1114.235716] env[62965]: DEBUG nova.network.neutron [req-03052458-5ecf-41c0-8fcd-f9472d19f171 req-25c33daf-e106-48af-9810-b5e3fd6ab141 service nova] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Updated VIF entry in instance network info cache for port 5f939a36-bc5f-4fd7-b29e-011f72b3256c. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1114.235938] env[62965]: DEBUG nova.network.neutron [req-03052458-5ecf-41c0-8fcd-f9472d19f171 req-25c33daf-e106-48af-9810-b5e3fd6ab141 service nova] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Updating instance_info_cache with network_info: [{"id": "5f939a36-bc5f-4fd7-b29e-011f72b3256c", "address": "fa:16:3e:e3:06:99", "network": {"id": "49dd42b0-6397-4fdd-a319-ed920c72a5af", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-373172566-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d91c99b14018469e8998fd1ccc0fab4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap5f939a36-bc", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1114.270604] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 4fc543a1-34ce-40bb-9e80-72ab9098eb38] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1114.525021] env[62965]: DEBUG nova.compute.manager [None req-06f31f85-a32b-46ae-bfb9-427281418d85 tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1114.525021] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-06f31f85-a32b-46ae-bfb9-427281418d85 tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1114.525021] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9192b5ff-ecdb-4bec-b5ff-1d244b0a89d2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.532269] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-06f31f85-a32b-46ae-bfb9-427281418d85 tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1114.532764] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-473a8f55-07cc-43b5-ad85-ef2250c813d8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.539405] env[62965]: DEBUG oslo_vmware.api [None req-06f31f85-a32b-46ae-bfb9-427281418d85 tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Waiting for the task: (returnval){ [ 1114.539405] env[62965]: value = "task-1390420" [ 1114.539405] env[62965]: _type = "Task" [ 1114.539405] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.547096] env[62965]: DEBUG oslo_vmware.api [None req-06f31f85-a32b-46ae-bfb9-427281418d85 tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Task: {'id': task-1390420, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.732054] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.127s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.733931] env[62965]: DEBUG oslo_concurrency.lockutils [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.510s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.734150] env[62965]: DEBUG oslo_concurrency.lockutils [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.737807] env[62965]: DEBUG oslo_concurrency.lockutils [req-03052458-5ecf-41c0-8fcd-f9472d19f171 req-25c33daf-e106-48af-9810-b5e3fd6ab141 service nova] Releasing lock "refresh_cache-73a7ca0e-b1d3-4464-969b-5111d2506a8b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1114.754644] env[62965]: INFO nova.scheduler.client.report [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Deleted allocations for instance b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7 [ 1114.773803] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: a0e9cc1a-bc95-4a44-99c5-aabf85a373f4] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1114.920517] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquiring lock "73a7ca0e-b1d3-4464-969b-5111d2506a8b" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1115.049527] env[62965]: DEBUG oslo_vmware.api [None req-06f31f85-a32b-46ae-bfb9-427281418d85 tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Task: {'id': task-1390420, 'name': PowerOffVM_Task, 'duration_secs': 0.194737} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.049809] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-06f31f85-a32b-46ae-bfb9-427281418d85 tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1115.049982] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-06f31f85-a32b-46ae-bfb9-427281418d85 tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1115.050266] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-642a5bfd-e8f1-4e8a-a71a-2eb4ea2c9fbb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.117601] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-06f31f85-a32b-46ae-bfb9-427281418d85 tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1115.117897] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-06f31f85-a32b-46ae-bfb9-427281418d85 tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Deleting contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1115.118135] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-06f31f85-a32b-46ae-bfb9-427281418d85 tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Deleting the datastore file [datastore2] 924c5cdc-0450-4ced-b920-50e2d5060fd2 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1115.118467] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3d9afe19-b288-4032-bd03-02463878bc7f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.126837] env[62965]: DEBUG oslo_vmware.api [None req-06f31f85-a32b-46ae-bfb9-427281418d85 tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Waiting for the task: (returnval){ [ 1115.126837] env[62965]: value = "task-1390423" [ 1115.126837] env[62965]: _type = "Task" [ 1115.126837] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.134489] env[62965]: DEBUG oslo_vmware.api [None req-06f31f85-a32b-46ae-bfb9-427281418d85 tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Task: {'id': task-1390423, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.241639] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ee6b68a1-1cf4-409a-b2a2-048aade211ad tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "73a7ca0e-b1d3-4464-969b-5111d2506a8b" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 21.006s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.242603] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "73a7ca0e-b1d3-4464-969b-5111d2506a8b" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.322s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.242828] env[62965]: INFO nova.compute.manager [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Unshelving [ 1115.263464] env[62965]: DEBUG oslo_concurrency.lockutils [None req-dba57953-c6ec-4618-b825-ca620f23286a tempest-DeleteServersTestJSON-980160634 tempest-DeleteServersTestJSON-980160634-project-member] Lock "b5f8f8ed-7996-49c1-a5cd-7a8a7f8907d7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.423s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.278061] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: bd36601b-5a70-4a29-8ef1-d1e925f41de7] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1115.635159] env[62965]: DEBUG oslo_vmware.api [None req-06f31f85-a32b-46ae-bfb9-427281418d85 tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Task: {'id': task-1390423, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136954} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.635458] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-06f31f85-a32b-46ae-bfb9-427281418d85 tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1115.635660] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-06f31f85-a32b-46ae-bfb9-427281418d85 tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Deleted contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1115.635839] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-06f31f85-a32b-46ae-bfb9-427281418d85 tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1115.636026] env[62965]: INFO nova.compute.manager [None req-06f31f85-a32b-46ae-bfb9-427281418d85 tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1115.636273] env[62965]: DEBUG oslo.service.loopingcall [None req-06f31f85-a32b-46ae-bfb9-427281418d85 tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1115.636478] env[62965]: DEBUG nova.compute.manager [-] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1115.636574] env[62965]: DEBUG nova.network.neutron [-] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1115.780925] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: b59d6023-a047-4b32-bd18-1c7cc05f67d7] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1116.093875] env[62965]: DEBUG nova.compute.manager [req-6f54d419-a636-4a0d-8860-f46bacf43935 req-7f09ef4f-a27b-4964-83a9-6bc16a6f2051 service nova] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Received event network-vif-deleted-80afc0d2-0d83-4374-a4c0-792eabf66c40 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1116.094108] env[62965]: INFO nova.compute.manager [req-6f54d419-a636-4a0d-8860-f46bacf43935 req-7f09ef4f-a27b-4964-83a9-6bc16a6f2051 service nova] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Neutron deleted interface 80afc0d2-0d83-4374-a4c0-792eabf66c40; detaching it from the instance and deleting it from the info cache [ 1116.094283] env[62965]: DEBUG nova.network.neutron [req-6f54d419-a636-4a0d-8860-f46bacf43935 req-7f09ef4f-a27b-4964-83a9-6bc16a6f2051 service nova] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1116.269443] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.269917] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.269917] env[62965]: DEBUG nova.objects.instance [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lazy-loading 'pci_requests' on Instance uuid 73a7ca0e-b1d3-4464-969b-5111d2506a8b {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1116.284943] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 6def4a38-d3b8-4724-97a1-1859b239d900] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1116.570790] env[62965]: DEBUG nova.network.neutron [-] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1116.597161] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-234e188b-7765-4738-9a98-f9d4a0b71837 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.606825] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7f78911-3108-4e5d-8e74-69e1435c3519 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.630672] env[62965]: DEBUG nova.compute.manager [req-6f54d419-a636-4a0d-8860-f46bacf43935 req-7f09ef4f-a27b-4964-83a9-6bc16a6f2051 service nova] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Detach interface failed, port_id=80afc0d2-0d83-4374-a4c0-792eabf66c40, reason: Instance 924c5cdc-0450-4ced-b920-50e2d5060fd2 could not be found. {{(pid=62965) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11294}} [ 1116.774178] env[62965]: DEBUG nova.objects.instance [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lazy-loading 'numa_topology' on Instance uuid 73a7ca0e-b1d3-4464-969b-5111d2506a8b {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1116.788253] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 8c478341-9b50-45de-9ba3-321e802ed1f0] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1117.073716] env[62965]: INFO nova.compute.manager [-] [instance: 924c5cdc-0450-4ced-b920-50e2d5060fd2] Took 1.44 seconds to deallocate network for instance. [ 1117.276950] env[62965]: INFO nova.compute.claims [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1117.291386] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: b0d8f38a-86d3-466a-9b83-35145071dc21] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1117.579862] env[62965]: DEBUG oslo_concurrency.lockutils [None req-06f31f85-a32b-46ae-bfb9-427281418d85 tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.794125] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: d4d9b202-3aa7-4b66-949d-4a2eb34b3adf] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1118.296824] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: c2448b76-1553-4b68-a731-0bb0967f4c1d] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1118.341180] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02afb766-2459-4c94-a7a8-1a61001bf639 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.348778] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92ed6407-0965-40a8-9301-2d137141e905 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.379020] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7558beba-801b-4098-b0fe-73b5458e60e8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.386925] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4124b13-40dc-4efa-9a04-c5b3bbe21b01 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.400856] env[62965]: DEBUG nova.compute.provider_tree [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1118.801328] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: d769d4ad-7d43-47a0-a0ba-abf3a93aaf8c] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1118.904150] env[62965]: DEBUG nova.scheduler.client.report [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1119.304737] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 64cf898c-2c97-4daa-80c8-0a1c252cbcf2] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1119.409516] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.140s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1119.412742] env[62965]: DEBUG oslo_concurrency.lockutils [None req-06f31f85-a32b-46ae-bfb9-427281418d85 tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.832s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.413976] env[62965]: DEBUG nova.objects.instance [None req-06f31f85-a32b-46ae-bfb9-427281418d85 tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Lazy-loading 'resources' on Instance uuid 924c5cdc-0450-4ced-b920-50e2d5060fd2 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1119.448026] env[62965]: INFO nova.network.neutron [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Updating port 5f939a36-bc5f-4fd7-b29e-011f72b3256c with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1119.808012] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: e87adfae-3ea4-4cc4-9978-d8c811c80941] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1119.968982] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-363bfa38-3afc-4760-bf2b-0ec9ef3d0407 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.976949] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81083add-97eb-4c70-9b4c-2b61adba7075 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.006531] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b77afaa4-98d9-460d-a1f7-446c4d3f0913 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.013545] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e875bfe-e681-44f2-9306-08a6baa3422c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.026048] env[62965]: DEBUG nova.compute.provider_tree [None req-06f31f85-a32b-46ae-bfb9-427281418d85 tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1120.311576] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 6e0a4e4b-21a2-4185-999e-69b4c74c37d7] Instance has had 0 of 5 cleanup attempts {{(pid=62965) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11570}} [ 1120.528913] env[62965]: DEBUG nova.scheduler.client.report [None req-06f31f85-a32b-46ae-bfb9-427281418d85 tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1120.816029] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1120.816205] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Cleaning up deleted instances with incomplete migration {{(pid=62965) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11595}} [ 1120.926645] env[62965]: DEBUG nova.compute.manager [req-cd273102-b1ff-4b72-ac5d-d53516ee7d18 req-6989281e-529e-4410-82b0-a69a41ba4623 service nova] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Received event network-vif-plugged-5f939a36-bc5f-4fd7-b29e-011f72b3256c {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1120.926868] env[62965]: DEBUG oslo_concurrency.lockutils [req-cd273102-b1ff-4b72-ac5d-d53516ee7d18 req-6989281e-529e-4410-82b0-a69a41ba4623 service nova] Acquiring lock "73a7ca0e-b1d3-4464-969b-5111d2506a8b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.927115] env[62965]: DEBUG oslo_concurrency.lockutils [req-cd273102-b1ff-4b72-ac5d-d53516ee7d18 req-6989281e-529e-4410-82b0-a69a41ba4623 service nova] Lock "73a7ca0e-b1d3-4464-969b-5111d2506a8b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1120.927257] env[62965]: DEBUG oslo_concurrency.lockutils [req-cd273102-b1ff-4b72-ac5d-d53516ee7d18 req-6989281e-529e-4410-82b0-a69a41ba4623 service nova] Lock "73a7ca0e-b1d3-4464-969b-5111d2506a8b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1120.927427] env[62965]: DEBUG nova.compute.manager [req-cd273102-b1ff-4b72-ac5d-d53516ee7d18 req-6989281e-529e-4410-82b0-a69a41ba4623 service nova] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] No waiting events found dispatching network-vif-plugged-5f939a36-bc5f-4fd7-b29e-011f72b3256c {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1120.927595] env[62965]: WARNING nova.compute.manager [req-cd273102-b1ff-4b72-ac5d-d53516ee7d18 req-6989281e-529e-4410-82b0-a69a41ba4623 service nova] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Received unexpected event network-vif-plugged-5f939a36-bc5f-4fd7-b29e-011f72b3256c for instance with vm_state shelved_offloaded and task_state spawning. [ 1121.013790] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquiring lock "refresh_cache-73a7ca0e-b1d3-4464-969b-5111d2506a8b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1121.014120] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquired lock "refresh_cache-73a7ca0e-b1d3-4464-969b-5111d2506a8b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1121.014314] env[62965]: DEBUG nova.network.neutron [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1121.033231] env[62965]: DEBUG oslo_concurrency.lockutils [None req-06f31f85-a32b-46ae-bfb9-427281418d85 tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.621s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.053592] env[62965]: INFO nova.scheduler.client.report [None req-06f31f85-a32b-46ae-bfb9-427281418d85 tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Deleted allocations for instance 924c5cdc-0450-4ced-b920-50e2d5060fd2 [ 1121.319206] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1121.560874] env[62965]: DEBUG oslo_concurrency.lockutils [None req-06f31f85-a32b-46ae-bfb9-427281418d85 tempest-ServersV294TestFqdnHostnames-917528896 tempest-ServersV294TestFqdnHostnames-917528896-project-member] Lock "924c5cdc-0450-4ced-b920-50e2d5060fd2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.544s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.738744] env[62965]: DEBUG nova.network.neutron [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Updating instance_info_cache with network_info: [{"id": "5f939a36-bc5f-4fd7-b29e-011f72b3256c", "address": "fa:16:3e:e3:06:99", "network": {"id": "49dd42b0-6397-4fdd-a319-ed920c72a5af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-373172566-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d91c99b14018469e8998fd1ccc0fab4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f939a36-bc", "ovs_interfaceid": "5f939a36-bc5f-4fd7-b29e-011f72b3256c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1122.241214] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Releasing lock "refresh_cache-73a7ca0e-b1d3-4464-969b-5111d2506a8b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1122.268617] env[62965]: DEBUG nova.virt.hardware [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='eb46d52013ef6b8e600a2302cddd6838',container_format='bare',created_at=2024-10-31T12:40:21Z,direct_url=,disk_format='vmdk',id=f1265e9a-63c3-4300-9384-70985cea3f2f,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-423022810-shelved',owner='d91c99b14018469e8998fd1ccc0fab4f',properties=ImageMetaProps,protected=,size=31664640,status='active',tags=,updated_at=2024-10-31T12:40:35Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1122.269074] env[62965]: DEBUG nova.virt.hardware [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1122.269250] env[62965]: DEBUG nova.virt.hardware [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1122.269446] env[62965]: DEBUG nova.virt.hardware [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1122.269774] env[62965]: DEBUG nova.virt.hardware [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1122.269982] env[62965]: DEBUG nova.virt.hardware [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1122.270233] env[62965]: DEBUG nova.virt.hardware [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1122.270407] env[62965]: DEBUG nova.virt.hardware [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1122.270730] env[62965]: DEBUG nova.virt.hardware [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1122.270958] env[62965]: DEBUG nova.virt.hardware [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1122.271164] env[62965]: DEBUG nova.virt.hardware [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1122.272372] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfd4080d-e6d0-470f-96a7-bf03981b1468 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.281482] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-112d6ac8-c3fb-4fe2-af55-bb6739c4e7c0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.294648] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e3:06:99', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f77ff7a1-209c-4f3f-b2a0-fd817741e739', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5f939a36-bc5f-4fd7-b29e-011f72b3256c', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1122.301835] env[62965]: DEBUG oslo.service.loopingcall [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1122.302074] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1122.302285] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-58f5c348-899c-41fc-9d2a-cdb67ccb030a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.321141] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1122.321141] env[62965]: value = "task-1390424" [ 1122.321141] env[62965]: _type = "Task" [ 1122.321141] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.329363] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390424, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.835772] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390424, 'name': CreateVM_Task, 'duration_secs': 0.301523} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.836457] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1122.837805] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f1265e9a-63c3-4300-9384-70985cea3f2f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1122.837805] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f1265e9a-63c3-4300-9384-70985cea3f2f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1122.837805] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/f1265e9a-63c3-4300-9384-70985cea3f2f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1122.837805] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c838d195-5a36-4441-9818-490b45d48450 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.842458] env[62965]: DEBUG oslo_vmware.api [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 1122.842458] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]527a4b75-9cd0-44e1-31a3-d7e8e6364135" [ 1122.842458] env[62965]: _type = "Task" [ 1122.842458] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.861745] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f1265e9a-63c3-4300-9384-70985cea3f2f" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1122.861990] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Processing image f1265e9a-63c3-4300-9384-70985cea3f2f {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1122.862250] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/f1265e9a-63c3-4300-9384-70985cea3f2f/f1265e9a-63c3-4300-9384-70985cea3f2f.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1122.862433] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquired lock "[datastore2] devstack-image-cache_base/f1265e9a-63c3-4300-9384-70985cea3f2f/f1265e9a-63c3-4300-9384-70985cea3f2f.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1122.862620] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1122.862873] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-50e162a9-dd80-46a6-8fee-9dad1be1b8e5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.870794] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1122.870975] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1122.871660] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75426fad-bd9d-461e-86b0-e0bc08fd1fbc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.876208] env[62965]: DEBUG oslo_vmware.api [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 1122.876208] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52041506-d4f1-c4b3-eb48-79531a413fe9" [ 1122.876208] env[62965]: _type = "Task" [ 1122.876208] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.887993] env[62965]: DEBUG oslo_vmware.api [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52041506-d4f1-c4b3-eb48-79531a413fe9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.959348] env[62965]: DEBUG nova.compute.manager [req-cf684d8b-4e81-47c7-b987-15c8930eb401 req-fc8f1cb5-85f1-490b-a61e-d78e4d082e4b service nova] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Received event network-changed-5f939a36-bc5f-4fd7-b29e-011f72b3256c {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1122.959543] env[62965]: DEBUG nova.compute.manager [req-cf684d8b-4e81-47c7-b987-15c8930eb401 req-fc8f1cb5-85f1-490b-a61e-d78e4d082e4b service nova] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Refreshing instance network info cache due to event network-changed-5f939a36-bc5f-4fd7-b29e-011f72b3256c. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 1122.960199] env[62965]: DEBUG oslo_concurrency.lockutils [req-cf684d8b-4e81-47c7-b987-15c8930eb401 req-fc8f1cb5-85f1-490b-a61e-d78e4d082e4b service nova] Acquiring lock "refresh_cache-73a7ca0e-b1d3-4464-969b-5111d2506a8b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1122.960199] env[62965]: DEBUG oslo_concurrency.lockutils [req-cf684d8b-4e81-47c7-b987-15c8930eb401 req-fc8f1cb5-85f1-490b-a61e-d78e4d082e4b service nova] Acquired lock "refresh_cache-73a7ca0e-b1d3-4464-969b-5111d2506a8b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1122.960309] env[62965]: DEBUG nova.network.neutron [req-cf684d8b-4e81-47c7-b987-15c8930eb401 req-fc8f1cb5-85f1-490b-a61e-d78e4d082e4b service nova] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Refreshing network info cache for port 5f939a36-bc5f-4fd7-b29e-011f72b3256c {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1123.387010] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Preparing fetch location {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1123.387309] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Fetch image to [datastore2] OSTACK_IMG_a3ba9e4b-346a-40cc-b134-110b64085c20/OSTACK_IMG_a3ba9e4b-346a-40cc-b134-110b64085c20.vmdk {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1123.387502] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Downloading stream optimized image f1265e9a-63c3-4300-9384-70985cea3f2f to [datastore2] OSTACK_IMG_a3ba9e4b-346a-40cc-b134-110b64085c20/OSTACK_IMG_a3ba9e4b-346a-40cc-b134-110b64085c20.vmdk on the data store datastore2 as vApp {{(pid=62965) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1123.387678] env[62965]: DEBUG nova.virt.vmwareapi.images [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Downloading image file data f1265e9a-63c3-4300-9384-70985cea3f2f to the ESX as VM named 'OSTACK_IMG_a3ba9e4b-346a-40cc-b134-110b64085c20' {{(pid=62965) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1123.458170] env[62965]: DEBUG oslo_vmware.rw_handles [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1123.458170] env[62965]: value = "resgroup-9" [ 1123.458170] env[62965]: _type = "ResourcePool" [ 1123.458170] env[62965]: }. {{(pid=62965) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1123.458439] env[62965]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-7ae49c3d-f708-4220-bc9e-673f70ef07f6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.481770] env[62965]: DEBUG oslo_vmware.rw_handles [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lease: (returnval){ [ 1123.481770] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52fcf18f-0e53-89c5-31f0-e888c3c8f95f" [ 1123.481770] env[62965]: _type = "HttpNfcLease" [ 1123.481770] env[62965]: } obtained for vApp import into resource pool (val){ [ 1123.481770] env[62965]: value = "resgroup-9" [ 1123.481770] env[62965]: _type = "ResourcePool" [ 1123.481770] env[62965]: }. {{(pid=62965) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1123.482016] env[62965]: DEBUG oslo_vmware.api [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the lease: (returnval){ [ 1123.482016] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52fcf18f-0e53-89c5-31f0-e888c3c8f95f" [ 1123.482016] env[62965]: _type = "HttpNfcLease" [ 1123.482016] env[62965]: } to be ready. {{(pid=62965) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1123.488456] env[62965]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1123.488456] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52fcf18f-0e53-89c5-31f0-e888c3c8f95f" [ 1123.488456] env[62965]: _type = "HttpNfcLease" [ 1123.488456] env[62965]: } is initializing. {{(pid=62965) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1123.723244] env[62965]: DEBUG nova.network.neutron [req-cf684d8b-4e81-47c7-b987-15c8930eb401 req-fc8f1cb5-85f1-490b-a61e-d78e4d082e4b service nova] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Updated VIF entry in instance network info cache for port 5f939a36-bc5f-4fd7-b29e-011f72b3256c. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1123.723642] env[62965]: DEBUG nova.network.neutron [req-cf684d8b-4e81-47c7-b987-15c8930eb401 req-fc8f1cb5-85f1-490b-a61e-d78e4d082e4b service nova] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Updating instance_info_cache with network_info: [{"id": "5f939a36-bc5f-4fd7-b29e-011f72b3256c", "address": "fa:16:3e:e3:06:99", "network": {"id": "49dd42b0-6397-4fdd-a319-ed920c72a5af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-373172566-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d91c99b14018469e8998fd1ccc0fab4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f939a36-bc", "ovs_interfaceid": "5f939a36-bc5f-4fd7-b29e-011f72b3256c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1123.989850] env[62965]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1123.989850] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52fcf18f-0e53-89c5-31f0-e888c3c8f95f" [ 1123.989850] env[62965]: _type = "HttpNfcLease" [ 1123.989850] env[62965]: } is ready. {{(pid=62965) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1123.990170] env[62965]: DEBUG oslo_vmware.rw_handles [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1123.990170] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52fcf18f-0e53-89c5-31f0-e888c3c8f95f" [ 1123.990170] env[62965]: _type = "HttpNfcLease" [ 1123.990170] env[62965]: }. {{(pid=62965) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1123.991285] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3de4a93-c11f-43f3-81ae-dab528711c5e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.999336] env[62965]: DEBUG oslo_vmware.rw_handles [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a0bc63-7849-e005-f18d-b81b27ae13d6/disk-0.vmdk from lease info. {{(pid=62965) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1123.999519] env[62965]: DEBUG oslo_vmware.rw_handles [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Creating HTTP connection to write to file with size = 31664640 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a0bc63-7849-e005-f18d-b81b27ae13d6/disk-0.vmdk. {{(pid=62965) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1124.065024] env[62965]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e183c633-fa4a-4769-ba7e-0acd7a77e68e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.226372] env[62965]: DEBUG oslo_concurrency.lockutils [req-cf684d8b-4e81-47c7-b987-15c8930eb401 req-fc8f1cb5-85f1-490b-a61e-d78e4d082e4b service nova] Releasing lock "refresh_cache-73a7ca0e-b1d3-4464-969b-5111d2506a8b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1125.162714] env[62965]: DEBUG oslo_vmware.rw_handles [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Completed reading data from the image iterator. {{(pid=62965) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1125.163180] env[62965]: DEBUG oslo_vmware.rw_handles [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a0bc63-7849-e005-f18d-b81b27ae13d6/disk-0.vmdk. {{(pid=62965) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1125.164086] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66e11a69-8111-4bde-b94d-9ee0defc200f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.171854] env[62965]: DEBUG oslo_vmware.rw_handles [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a0bc63-7849-e005-f18d-b81b27ae13d6/disk-0.vmdk is in state: ready. {{(pid=62965) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1125.172174] env[62965]: DEBUG oslo_vmware.rw_handles [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a0bc63-7849-e005-f18d-b81b27ae13d6/disk-0.vmdk. {{(pid=62965) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1125.172522] env[62965]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-94946c29-f301-4078-bfd8-4f591e5dd3bf {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.354814] env[62965]: DEBUG oslo_vmware.rw_handles [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52a0bc63-7849-e005-f18d-b81b27ae13d6/disk-0.vmdk. {{(pid=62965) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1125.355090] env[62965]: INFO nova.virt.vmwareapi.images [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Downloaded image file data f1265e9a-63c3-4300-9384-70985cea3f2f [ 1125.355953] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24fbbba2-9786-4bf8-b51a-ea801489955e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.373568] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f851a8d7-4f3c-4b33-b92d-b4353d678eab {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.398255] env[62965]: INFO nova.virt.vmwareapi.images [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] The imported VM was unregistered [ 1125.400756] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Caching image {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1125.400997] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Creating directory with path [datastore2] devstack-image-cache_base/f1265e9a-63c3-4300-9384-70985cea3f2f {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1125.401313] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8e6459f5-d32b-4afa-af04-efd122824a9f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.420291] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Created directory with path [datastore2] devstack-image-cache_base/f1265e9a-63c3-4300-9384-70985cea3f2f {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1125.420491] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_a3ba9e4b-346a-40cc-b134-110b64085c20/OSTACK_IMG_a3ba9e4b-346a-40cc-b134-110b64085c20.vmdk to [datastore2] devstack-image-cache_base/f1265e9a-63c3-4300-9384-70985cea3f2f/f1265e9a-63c3-4300-9384-70985cea3f2f.vmdk. {{(pid=62965) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1125.420772] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-cd253a6e-66af-4442-9647-a33d7e6659a3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.428384] env[62965]: DEBUG oslo_vmware.api [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 1125.428384] env[62965]: value = "task-1390427" [ 1125.428384] env[62965]: _type = "Task" [ 1125.428384] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.435639] env[62965]: DEBUG oslo_vmware.api [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390427, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.486761] env[62965]: DEBUG nova.compute.manager [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Stashing vm_state: active {{(pid=62965) _prep_resize /opt/stack/nova/nova/compute/manager.py:5953}} [ 1125.940480] env[62965]: DEBUG oslo_vmware.api [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390427, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.008782] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1126.009111] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1126.442243] env[62965]: DEBUG oslo_vmware.api [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390427, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.514525] env[62965]: INFO nova.compute.claims [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1126.940839] env[62965]: DEBUG oslo_vmware.api [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390427, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.024933] env[62965]: INFO nova.compute.resource_tracker [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Updating resource usage from migration ef5466ea-156c-429a-ae78-424d5f9443a8 [ 1127.095178] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-566819b5-0a61-4af4-aad1-f41b04a5ce49 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.105366] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96076628-d6bb-4a66-9479-26175fdf9c91 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.140686] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69697a0c-63f7-4edb-aca1-5cd06f93200d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.152476] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf377890-e75e-455b-9573-8f45907ae8cc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.170806] env[62965]: DEBUG nova.compute.provider_tree [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1127.215251] env[62965]: DEBUG oslo_concurrency.lockutils [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Acquiring lock "e8ff0b40-b329-404f-a0c8-f123c1fd6d40" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.215514] env[62965]: DEBUG oslo_concurrency.lockutils [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Lock "e8ff0b40-b329-404f-a0c8-f123c1fd6d40" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.441341] env[62965]: DEBUG oslo_vmware.api [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390427, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.675212] env[62965]: DEBUG nova.scheduler.client.report [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1127.718101] env[62965]: DEBUG nova.compute.manager [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1127.940540] env[62965]: DEBUG oslo_vmware.api [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390427, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.376157} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.940839] env[62965]: INFO nova.virt.vmwareapi.ds_util [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_a3ba9e4b-346a-40cc-b134-110b64085c20/OSTACK_IMG_a3ba9e4b-346a-40cc-b134-110b64085c20.vmdk to [datastore2] devstack-image-cache_base/f1265e9a-63c3-4300-9384-70985cea3f2f/f1265e9a-63c3-4300-9384-70985cea3f2f.vmdk. [ 1127.941042] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Cleaning up location [datastore2] OSTACK_IMG_a3ba9e4b-346a-40cc-b134-110b64085c20 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1127.941213] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_a3ba9e4b-346a-40cc-b134-110b64085c20 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1127.941474] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fb4fe340-123d-4d21-b9d2-8823ac4adff0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.947712] env[62965]: DEBUG oslo_vmware.api [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 1127.947712] env[62965]: value = "task-1390428" [ 1127.947712] env[62965]: _type = "Task" [ 1127.947712] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.955157] env[62965]: DEBUG oslo_vmware.api [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390428, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.186021] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.176s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1128.186021] env[62965]: INFO nova.compute.manager [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Migrating [ 1128.242695] env[62965]: DEBUG oslo_concurrency.lockutils [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1128.243012] env[62965]: DEBUG oslo_concurrency.lockutils [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1128.244551] env[62965]: INFO nova.compute.claims [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1128.459629] env[62965]: DEBUG oslo_vmware.api [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390428, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.031399} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.460246] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1128.460591] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Releasing lock "[datastore2] devstack-image-cache_base/f1265e9a-63c3-4300-9384-70985cea3f2f/f1265e9a-63c3-4300-9384-70985cea3f2f.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1128.461020] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/f1265e9a-63c3-4300-9384-70985cea3f2f/f1265e9a-63c3-4300-9384-70985cea3f2f.vmdk to [datastore2] 73a7ca0e-b1d3-4464-969b-5111d2506a8b/73a7ca0e-b1d3-4464-969b-5111d2506a8b.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1128.461309] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1aa4306e-57a4-41c1-b8f2-a937e50d997e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.468051] env[62965]: DEBUG oslo_vmware.api [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 1128.468051] env[62965]: value = "task-1390429" [ 1128.468051] env[62965]: _type = "Task" [ 1128.468051] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.476017] env[62965]: DEBUG oslo_vmware.api [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390429, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.699674] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "refresh_cache-385648d7-20e3-442c-b6a1-3d972d266487" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1128.700051] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquired lock "refresh_cache-385648d7-20e3-442c-b6a1-3d972d266487" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1128.700051] env[62965]: DEBUG nova.network.neutron [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1128.980851] env[62965]: DEBUG oslo_vmware.api [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390429, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.316477] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4849bd99-de9a-455d-8cfe-40df5223b680 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.327663] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-604bcae6-4da8-4ae2-a3c8-d428f45fa172 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.365972] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ca07f9d-4627-4c41-b5c1-1e6383e14f47 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.375344] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b541b52-7543-4d25-955f-0c0b03a9ca45 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.391872] env[62965]: DEBUG nova.compute.provider_tree [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1129.438837] env[62965]: DEBUG nova.network.neutron [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Updating instance_info_cache with network_info: [{"id": "7de6b63c-f2c3-4a0b-b399-73075c5e09a5", "address": "fa:16:3e:6c:49:73", "network": {"id": "4649fed7-c2ca-43a9-a183-7c99423da726", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1978698173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fab418e2fe1420793517663574b43bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43ad01d2-c7dd-453c-a929-8ad76294d13c", "external-id": "nsx-vlan-transportzone-176", "segmentation_id": 176, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7de6b63c-f2", "ovs_interfaceid": "7de6b63c-f2c3-4a0b-b399-73075c5e09a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1129.483160] env[62965]: DEBUG oslo_vmware.api [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390429, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.895269] env[62965]: DEBUG nova.scheduler.client.report [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1129.941544] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Releasing lock "refresh_cache-385648d7-20e3-442c-b6a1-3d972d266487" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1129.982602] env[62965]: DEBUG oslo_vmware.api [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390429, 'name': CopyVirtualDisk_Task} progress is 66%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.400172] env[62965]: DEBUG oslo_concurrency.lockutils [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.157s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1130.400717] env[62965]: DEBUG nova.compute.manager [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1130.484206] env[62965]: DEBUG oslo_vmware.api [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390429, 'name': CopyVirtualDisk_Task} progress is 88%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.907038] env[62965]: DEBUG nova.compute.utils [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1130.908413] env[62965]: DEBUG nova.compute.manager [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1130.908582] env[62965]: DEBUG nova.network.neutron [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1130.954952] env[62965]: DEBUG nova.policy [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '777f044a91c14b5ca3f92b12a9569f94', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a757887b3e1e4c69bea03f4924fc813e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 1130.981662] env[62965]: DEBUG oslo_vmware.api [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390429, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.227849} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.981934] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/f1265e9a-63c3-4300-9384-70985cea3f2f/f1265e9a-63c3-4300-9384-70985cea3f2f.vmdk to [datastore2] 73a7ca0e-b1d3-4464-969b-5111d2506a8b/73a7ca0e-b1d3-4464-969b-5111d2506a8b.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1130.982752] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df9a84ce-0bd6-4400-a854-92c4fbc4d1b5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.004878] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 73a7ca0e-b1d3-4464-969b-5111d2506a8b/73a7ca0e-b1d3-4464-969b-5111d2506a8b.vmdk or device None with type streamOptimized {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1131.005382] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f8c12748-da9c-4b54-a1c5-7e9296f511cb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.023801] env[62965]: DEBUG oslo_vmware.api [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 1131.023801] env[62965]: value = "task-1390430" [ 1131.023801] env[62965]: _type = "Task" [ 1131.023801] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.031173] env[62965]: DEBUG oslo_vmware.api [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390430, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.185022] env[62965]: DEBUG nova.network.neutron [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Successfully created port: 1f01a2f6-4965-4721-a5b4-7d1c4b1fde60 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1131.403412] env[62965]: DEBUG nova.network.neutron [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Successfully created port: 84a65364-a7f3-4ed6-8496-44a1e968fe12 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1131.412640] env[62965]: DEBUG nova.compute.manager [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1131.457773] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdae8500-3a54-4668-af6b-ae9dba00011f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.478931] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Updating instance '385648d7-20e3-442c-b6a1-3d972d266487' progress to 0 {{(pid=62965) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1131.534302] env[62965]: DEBUG oslo_vmware.api [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390430, 'name': ReconfigVM_Task, 'duration_secs': 0.26616} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.534594] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 73a7ca0e-b1d3-4464-969b-5111d2506a8b/73a7ca0e-b1d3-4464-969b-5111d2506a8b.vmdk or device None with type streamOptimized {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1131.535282] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-086a82d8-e7cc-41e2-9e02-de6b40fdef6c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.541952] env[62965]: DEBUG oslo_vmware.api [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 1131.541952] env[62965]: value = "task-1390431" [ 1131.541952] env[62965]: _type = "Task" [ 1131.541952] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.550401] env[62965]: DEBUG oslo_vmware.api [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390431, 'name': Rename_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.641245] env[62965]: DEBUG nova.network.neutron [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Successfully created port: fd004411-8e83-4378-b85c-098463c3e810 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1131.986044] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1131.986044] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d68cc47d-0675-4d00-b2c4-d225c3531436 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.993621] env[62965]: DEBUG oslo_vmware.api [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1131.993621] env[62965]: value = "task-1390432" [ 1131.993621] env[62965]: _type = "Task" [ 1131.993621] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.002640] env[62965]: DEBUG oslo_vmware.api [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390432, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.056455] env[62965]: DEBUG oslo_vmware.api [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390431, 'name': Rename_Task} progress is 99%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.422922] env[62965]: DEBUG nova.compute.manager [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1132.450768] env[62965]: DEBUG nova.virt.hardware [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1132.451201] env[62965]: DEBUG nova.virt.hardware [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1132.451795] env[62965]: DEBUG nova.virt.hardware [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1132.451795] env[62965]: DEBUG nova.virt.hardware [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1132.451999] env[62965]: DEBUG nova.virt.hardware [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1132.452267] env[62965]: DEBUG nova.virt.hardware [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1132.452700] env[62965]: DEBUG nova.virt.hardware [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1132.452921] env[62965]: DEBUG nova.virt.hardware [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1132.453130] env[62965]: DEBUG nova.virt.hardware [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1132.453312] env[62965]: DEBUG nova.virt.hardware [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1132.453499] env[62965]: DEBUG nova.virt.hardware [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1132.454572] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dd55a91-9b9f-4a4b-bf0b-a47b92ff213d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.463251] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7ffdca7-586d-463b-ab60-cf471b73d203 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.502616] env[62965]: DEBUG oslo_vmware.api [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390432, 'name': PowerOffVM_Task, 'duration_secs': 0.400035} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.502906] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1132.503117] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Updating instance '385648d7-20e3-442c-b6a1-3d972d266487' progress to 17 {{(pid=62965) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1132.553135] env[62965]: DEBUG oslo_vmware.api [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390431, 'name': Rename_Task, 'duration_secs': 0.524261} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.553441] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1132.553820] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-903e28cd-a17f-4575-95cd-4867fcaa7593 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.560881] env[62965]: DEBUG oslo_vmware.api [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 1132.560881] env[62965]: value = "task-1390433" [ 1132.560881] env[62965]: _type = "Task" [ 1132.560881] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.568978] env[62965]: DEBUG oslo_vmware.api [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390433, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.009607] env[62965]: DEBUG nova.virt.hardware [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:24Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1133.009982] env[62965]: DEBUG nova.virt.hardware [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1133.009982] env[62965]: DEBUG nova.virt.hardware [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1133.010135] env[62965]: DEBUG nova.virt.hardware [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1133.010212] env[62965]: DEBUG nova.virt.hardware [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1133.010370] env[62965]: DEBUG nova.virt.hardware [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1133.010622] env[62965]: DEBUG nova.virt.hardware [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1133.010812] env[62965]: DEBUG nova.virt.hardware [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1133.010988] env[62965]: DEBUG nova.virt.hardware [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1133.011425] env[62965]: DEBUG nova.virt.hardware [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1133.011742] env[62965]: DEBUG nova.virt.hardware [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1133.017014] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-42c0cfb1-71bf-45aa-9d54-42366ddb9952 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.034473] env[62965]: DEBUG oslo_vmware.api [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1133.034473] env[62965]: value = "task-1390434" [ 1133.034473] env[62965]: _type = "Task" [ 1133.034473] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.040273] env[62965]: DEBUG nova.compute.manager [req-90004bd9-30c2-4c08-89f2-01868fc13282 req-6acc9c59-5ba3-40da-abbe-9c5cddf3edaa service nova] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Received event network-vif-plugged-1f01a2f6-4965-4721-a5b4-7d1c4b1fde60 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1133.040980] env[62965]: DEBUG oslo_concurrency.lockutils [req-90004bd9-30c2-4c08-89f2-01868fc13282 req-6acc9c59-5ba3-40da-abbe-9c5cddf3edaa service nova] Acquiring lock "e8ff0b40-b329-404f-a0c8-f123c1fd6d40-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1133.040980] env[62965]: DEBUG oslo_concurrency.lockutils [req-90004bd9-30c2-4c08-89f2-01868fc13282 req-6acc9c59-5ba3-40da-abbe-9c5cddf3edaa service nova] Lock "e8ff0b40-b329-404f-a0c8-f123c1fd6d40-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.041186] env[62965]: DEBUG oslo_concurrency.lockutils [req-90004bd9-30c2-4c08-89f2-01868fc13282 req-6acc9c59-5ba3-40da-abbe-9c5cddf3edaa service nova] Lock "e8ff0b40-b329-404f-a0c8-f123c1fd6d40-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.041484] env[62965]: DEBUG nova.compute.manager [req-90004bd9-30c2-4c08-89f2-01868fc13282 req-6acc9c59-5ba3-40da-abbe-9c5cddf3edaa service nova] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] No waiting events found dispatching network-vif-plugged-1f01a2f6-4965-4721-a5b4-7d1c4b1fde60 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1133.041705] env[62965]: WARNING nova.compute.manager [req-90004bd9-30c2-4c08-89f2-01868fc13282 req-6acc9c59-5ba3-40da-abbe-9c5cddf3edaa service nova] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Received unexpected event network-vif-plugged-1f01a2f6-4965-4721-a5b4-7d1c4b1fde60 for instance with vm_state building and task_state spawning. [ 1133.049725] env[62965]: DEBUG oslo_vmware.api [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390434, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.073050] env[62965]: DEBUG oslo_vmware.api [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390433, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.137976] env[62965]: DEBUG nova.network.neutron [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Successfully updated port: 1f01a2f6-4965-4721-a5b4-7d1c4b1fde60 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1133.544090] env[62965]: DEBUG oslo_vmware.api [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390434, 'name': ReconfigVM_Task, 'duration_secs': 0.157362} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.544400] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Updating instance '385648d7-20e3-442c-b6a1-3d972d266487' progress to 33 {{(pid=62965) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1133.570996] env[62965]: DEBUG oslo_vmware.api [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390433, 'name': PowerOnVM_Task, 'duration_secs': 0.555468} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.571272] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1133.664433] env[62965]: DEBUG nova.compute.manager [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1133.665381] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3b8f583-7232-40fd-8117-69d48e2db308 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.050825] env[62965]: DEBUG nova.virt.hardware [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1134.051239] env[62965]: DEBUG nova.virt.hardware [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1134.051239] env[62965]: DEBUG nova.virt.hardware [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1134.051377] env[62965]: DEBUG nova.virt.hardware [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1134.051528] env[62965]: DEBUG nova.virt.hardware [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1134.051678] env[62965]: DEBUG nova.virt.hardware [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1134.051885] env[62965]: DEBUG nova.virt.hardware [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1134.052061] env[62965]: DEBUG nova.virt.hardware [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1134.052241] env[62965]: DEBUG nova.virt.hardware [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1134.052406] env[62965]: DEBUG nova.virt.hardware [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1134.052612] env[62965]: DEBUG nova.virt.hardware [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1134.057805] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Reconfiguring VM instance instance-0000006a to detach disk 2000 {{(pid=62965) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1134.058116] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-00fa3788-07d4-4802-bce6-41fcdf39632e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.076643] env[62965]: DEBUG oslo_vmware.api [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1134.076643] env[62965]: value = "task-1390435" [ 1134.076643] env[62965]: _type = "Task" [ 1134.076643] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.084650] env[62965]: DEBUG oslo_vmware.api [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390435, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.184133] env[62965]: DEBUG oslo_concurrency.lockutils [None req-1c1029af-ab8a-4f12-9a82-d81c7f5abd4c tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "73a7ca0e-b1d3-4464-969b-5111d2506a8b" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 18.941s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1134.586320] env[62965]: DEBUG oslo_vmware.api [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390435, 'name': ReconfigVM_Task, 'duration_secs': 0.158273} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.586604] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Reconfigured VM instance instance-0000006a to detach disk 2000 {{(pid=62965) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1134.587437] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea7031c3-ef86-4dad-bc08-5d2dffda8df3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.608951] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] 385648d7-20e3-442c-b6a1-3d972d266487/385648d7-20e3-442c-b6a1-3d972d266487.vmdk or device None with type thin {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1134.609200] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-58f46ed9-c935-4d39-ac5c-73616bc0b324 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.627464] env[62965]: DEBUG oslo_vmware.api [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1134.627464] env[62965]: value = "task-1390436" [ 1134.627464] env[62965]: _type = "Task" [ 1134.627464] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.638639] env[62965]: DEBUG oslo_vmware.api [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390436, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.065761] env[62965]: DEBUG nova.compute.manager [req-2a6a5c9d-eee7-4dfa-a03e-d64a8e51dad1 req-cd09cdac-706a-4a84-846b-7b1ee94fef98 service nova] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Received event network-changed-1f01a2f6-4965-4721-a5b4-7d1c4b1fde60 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1135.065994] env[62965]: DEBUG nova.compute.manager [req-2a6a5c9d-eee7-4dfa-a03e-d64a8e51dad1 req-cd09cdac-706a-4a84-846b-7b1ee94fef98 service nova] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Refreshing instance network info cache due to event network-changed-1f01a2f6-4965-4721-a5b4-7d1c4b1fde60. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 1135.066316] env[62965]: DEBUG oslo_concurrency.lockutils [req-2a6a5c9d-eee7-4dfa-a03e-d64a8e51dad1 req-cd09cdac-706a-4a84-846b-7b1ee94fef98 service nova] Acquiring lock "refresh_cache-e8ff0b40-b329-404f-a0c8-f123c1fd6d40" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1135.066529] env[62965]: DEBUG oslo_concurrency.lockutils [req-2a6a5c9d-eee7-4dfa-a03e-d64a8e51dad1 req-cd09cdac-706a-4a84-846b-7b1ee94fef98 service nova] Acquired lock "refresh_cache-e8ff0b40-b329-404f-a0c8-f123c1fd6d40" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1135.066763] env[62965]: DEBUG nova.network.neutron [req-2a6a5c9d-eee7-4dfa-a03e-d64a8e51dad1 req-cd09cdac-706a-4a84-846b-7b1ee94fef98 service nova] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Refreshing network info cache for port 1f01a2f6-4965-4721-a5b4-7d1c4b1fde60 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1135.102892] env[62965]: DEBUG nova.network.neutron [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Successfully updated port: 84a65364-a7f3-4ed6-8496-44a1e968fe12 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1135.138617] env[62965]: DEBUG oslo_vmware.api [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390436, 'name': ReconfigVM_Task, 'duration_secs': 0.262254} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.138860] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Reconfigured VM instance instance-0000006a to attach disk [datastore1] 385648d7-20e3-442c-b6a1-3d972d266487/385648d7-20e3-442c-b6a1-3d972d266487.vmdk or device None with type thin {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1135.139104] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Updating instance '385648d7-20e3-442c-b6a1-3d972d266487' progress to 50 {{(pid=62965) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1135.608015] env[62965]: DEBUG nova.network.neutron [req-2a6a5c9d-eee7-4dfa-a03e-d64a8e51dad1 req-cd09cdac-706a-4a84-846b-7b1ee94fef98 service nova] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1135.645514] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54a3b817-0d0c-42a0-9c9b-fc7344fd7ba6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.667581] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c90026f8-e9c1-4a20-9075-bb65f3cfbe6c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.686346] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Updating instance '385648d7-20e3-442c-b6a1-3d972d266487' progress to 67 {{(pid=62965) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1135.709215] env[62965]: DEBUG nova.network.neutron [req-2a6a5c9d-eee7-4dfa-a03e-d64a8e51dad1 req-cd09cdac-706a-4a84-846b-7b1ee94fef98 service nova] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1136.211456] env[62965]: DEBUG oslo_concurrency.lockutils [req-2a6a5c9d-eee7-4dfa-a03e-d64a8e51dad1 req-cd09cdac-706a-4a84-846b-7b1ee94fef98 service nova] Releasing lock "refresh_cache-e8ff0b40-b329-404f-a0c8-f123c1fd6d40" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1136.211968] env[62965]: DEBUG nova.compute.manager [req-2a6a5c9d-eee7-4dfa-a03e-d64a8e51dad1 req-cd09cdac-706a-4a84-846b-7b1ee94fef98 service nova] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Received event network-vif-plugged-84a65364-a7f3-4ed6-8496-44a1e968fe12 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1136.211968] env[62965]: DEBUG oslo_concurrency.lockutils [req-2a6a5c9d-eee7-4dfa-a03e-d64a8e51dad1 req-cd09cdac-706a-4a84-846b-7b1ee94fef98 service nova] Acquiring lock "e8ff0b40-b329-404f-a0c8-f123c1fd6d40-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1136.212150] env[62965]: DEBUG oslo_concurrency.lockutils [req-2a6a5c9d-eee7-4dfa-a03e-d64a8e51dad1 req-cd09cdac-706a-4a84-846b-7b1ee94fef98 service nova] Lock "e8ff0b40-b329-404f-a0c8-f123c1fd6d40-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1136.212322] env[62965]: DEBUG oslo_concurrency.lockutils [req-2a6a5c9d-eee7-4dfa-a03e-d64a8e51dad1 req-cd09cdac-706a-4a84-846b-7b1ee94fef98 service nova] Lock "e8ff0b40-b329-404f-a0c8-f123c1fd6d40-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1136.212491] env[62965]: DEBUG nova.compute.manager [req-2a6a5c9d-eee7-4dfa-a03e-d64a8e51dad1 req-cd09cdac-706a-4a84-846b-7b1ee94fef98 service nova] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] No waiting events found dispatching network-vif-plugged-84a65364-a7f3-4ed6-8496-44a1e968fe12 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1136.212700] env[62965]: WARNING nova.compute.manager [req-2a6a5c9d-eee7-4dfa-a03e-d64a8e51dad1 req-cd09cdac-706a-4a84-846b-7b1ee94fef98 service nova] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Received unexpected event network-vif-plugged-84a65364-a7f3-4ed6-8496-44a1e968fe12 for instance with vm_state building and task_state spawning. [ 1136.224681] env[62965]: DEBUG nova.network.neutron [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Port 7de6b63c-f2c3-4a0b-b399-73075c5e09a5 binding to destination host cpu-1 is already ACTIVE {{(pid=62965) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1136.993358] env[62965]: DEBUG nova.compute.manager [req-e8f0ecd0-4a19-4247-a97f-72814eed603d req-55b1d6ca-4f24-4695-8091-7cbefc0503b2 service nova] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Received event network-vif-plugged-fd004411-8e83-4378-b85c-098463c3e810 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1136.993593] env[62965]: DEBUG oslo_concurrency.lockutils [req-e8f0ecd0-4a19-4247-a97f-72814eed603d req-55b1d6ca-4f24-4695-8091-7cbefc0503b2 service nova] Acquiring lock "e8ff0b40-b329-404f-a0c8-f123c1fd6d40-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1136.993835] env[62965]: DEBUG oslo_concurrency.lockutils [req-e8f0ecd0-4a19-4247-a97f-72814eed603d req-55b1d6ca-4f24-4695-8091-7cbefc0503b2 service nova] Lock "e8ff0b40-b329-404f-a0c8-f123c1fd6d40-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1136.994028] env[62965]: DEBUG oslo_concurrency.lockutils [req-e8f0ecd0-4a19-4247-a97f-72814eed603d req-55b1d6ca-4f24-4695-8091-7cbefc0503b2 service nova] Lock "e8ff0b40-b329-404f-a0c8-f123c1fd6d40-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1136.994905] env[62965]: DEBUG nova.compute.manager [req-e8f0ecd0-4a19-4247-a97f-72814eed603d req-55b1d6ca-4f24-4695-8091-7cbefc0503b2 service nova] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] No waiting events found dispatching network-vif-plugged-fd004411-8e83-4378-b85c-098463c3e810 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1136.995158] env[62965]: WARNING nova.compute.manager [req-e8f0ecd0-4a19-4247-a97f-72814eed603d req-55b1d6ca-4f24-4695-8091-7cbefc0503b2 service nova] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Received unexpected event network-vif-plugged-fd004411-8e83-4378-b85c-098463c3e810 for instance with vm_state building and task_state spawning. [ 1137.078558] env[62965]: DEBUG nova.network.neutron [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Successfully updated port: fd004411-8e83-4378-b85c-098463c3e810 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1137.092361] env[62965]: DEBUG nova.compute.manager [req-ec2453b9-0b2d-44d2-bcc6-435c8b42cabe req-7a1c0829-8dbd-4794-87ef-c25157b43cc2 service nova] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Received event network-changed-84a65364-a7f3-4ed6-8496-44a1e968fe12 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1137.092605] env[62965]: DEBUG nova.compute.manager [req-ec2453b9-0b2d-44d2-bcc6-435c8b42cabe req-7a1c0829-8dbd-4794-87ef-c25157b43cc2 service nova] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Refreshing instance network info cache due to event network-changed-84a65364-a7f3-4ed6-8496-44a1e968fe12. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 1137.093273] env[62965]: DEBUG oslo_concurrency.lockutils [req-ec2453b9-0b2d-44d2-bcc6-435c8b42cabe req-7a1c0829-8dbd-4794-87ef-c25157b43cc2 service nova] Acquiring lock "refresh_cache-e8ff0b40-b329-404f-a0c8-f123c1fd6d40" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1137.093273] env[62965]: DEBUG oslo_concurrency.lockutils [req-ec2453b9-0b2d-44d2-bcc6-435c8b42cabe req-7a1c0829-8dbd-4794-87ef-c25157b43cc2 service nova] Acquired lock "refresh_cache-e8ff0b40-b329-404f-a0c8-f123c1fd6d40" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1137.093273] env[62965]: DEBUG nova.network.neutron [req-ec2453b9-0b2d-44d2-bcc6-435c8b42cabe req-7a1c0829-8dbd-4794-87ef-c25157b43cc2 service nova] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Refreshing network info cache for port 84a65364-a7f3-4ed6-8496-44a1e968fe12 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1137.246466] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "385648d7-20e3-442c-b6a1-3d972d266487-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.246743] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "385648d7-20e3-442c-b6a1-3d972d266487-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.246858] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "385648d7-20e3-442c-b6a1-3d972d266487-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.583186] env[62965]: DEBUG oslo_concurrency.lockutils [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Acquiring lock "refresh_cache-e8ff0b40-b329-404f-a0c8-f123c1fd6d40" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1137.631155] env[62965]: DEBUG nova.network.neutron [req-ec2453b9-0b2d-44d2-bcc6-435c8b42cabe req-7a1c0829-8dbd-4794-87ef-c25157b43cc2 service nova] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1137.725330] env[62965]: DEBUG nova.network.neutron [req-ec2453b9-0b2d-44d2-bcc6-435c8b42cabe req-7a1c0829-8dbd-4794-87ef-c25157b43cc2 service nova] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1138.228476] env[62965]: DEBUG oslo_concurrency.lockutils [req-ec2453b9-0b2d-44d2-bcc6-435c8b42cabe req-7a1c0829-8dbd-4794-87ef-c25157b43cc2 service nova] Releasing lock "refresh_cache-e8ff0b40-b329-404f-a0c8-f123c1fd6d40" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1138.228874] env[62965]: DEBUG oslo_concurrency.lockutils [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Acquired lock "refresh_cache-e8ff0b40-b329-404f-a0c8-f123c1fd6d40" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1138.229062] env[62965]: DEBUG nova.network.neutron [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1138.278120] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "refresh_cache-385648d7-20e3-442c-b6a1-3d972d266487" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1138.278491] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquired lock "refresh_cache-385648d7-20e3-442c-b6a1-3d972d266487" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1138.278491] env[62965]: DEBUG nova.network.neutron [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1138.764083] env[62965]: DEBUG nova.network.neutron [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1138.981224] env[62965]: DEBUG nova.network.neutron [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Updating instance_info_cache with network_info: [{"id": "7de6b63c-f2c3-4a0b-b399-73075c5e09a5", "address": "fa:16:3e:6c:49:73", "network": {"id": "4649fed7-c2ca-43a9-a183-7c99423da726", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1978698173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fab418e2fe1420793517663574b43bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43ad01d2-c7dd-453c-a929-8ad76294d13c", "external-id": "nsx-vlan-transportzone-176", "segmentation_id": 176, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7de6b63c-f2", "ovs_interfaceid": "7de6b63c-f2c3-4a0b-b399-73075c5e09a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1139.019851] env[62965]: DEBUG nova.compute.manager [req-9705e850-51dd-400c-b747-4de901929061 req-e8fcc841-ce39-4aa0-962a-24f34549f388 service nova] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Received event network-changed-fd004411-8e83-4378-b85c-098463c3e810 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1139.020065] env[62965]: DEBUG nova.compute.manager [req-9705e850-51dd-400c-b747-4de901929061 req-e8fcc841-ce39-4aa0-962a-24f34549f388 service nova] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Refreshing instance network info cache due to event network-changed-fd004411-8e83-4378-b85c-098463c3e810. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 1139.020259] env[62965]: DEBUG oslo_concurrency.lockutils [req-9705e850-51dd-400c-b747-4de901929061 req-e8fcc841-ce39-4aa0-962a-24f34549f388 service nova] Acquiring lock "refresh_cache-e8ff0b40-b329-404f-a0c8-f123c1fd6d40" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1139.218033] env[62965]: DEBUG nova.network.neutron [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Updating instance_info_cache with network_info: [{"id": "1f01a2f6-4965-4721-a5b4-7d1c4b1fde60", "address": "fa:16:3e:51:d0:61", "network": {"id": "71361615-7a84-4c75-9481-afed7dc299d8", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2081071447", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.241", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a757887b3e1e4c69bea03f4924fc813e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c883fb98-d172-4510-8cf4-07aafdf771af", "external-id": "nsx-vlan-transportzone-570", "segmentation_id": 570, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f01a2f6-49", "ovs_interfaceid": "1f01a2f6-4965-4721-a5b4-7d1c4b1fde60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "84a65364-a7f3-4ed6-8496-44a1e968fe12", "address": "fa:16:3e:03:0e:62", "network": {"id": "d61ce9c6-768b-4f8b-8658-00e3b55893eb", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-949610621", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.63", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "a757887b3e1e4c69bea03f4924fc813e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37e8d2ee-abfc-42e2-a8fa-ee5447f1f1da", "external-id": "nsx-vlan-transportzone-813", "segmentation_id": 813, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap84a65364-a7", "ovs_interfaceid": "84a65364-a7f3-4ed6-8496-44a1e968fe12", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "fd004411-8e83-4378-b85c-098463c3e810", "address": "fa:16:3e:1e:0e:d2", "network": {"id": "71361615-7a84-4c75-9481-afed7dc299d8", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2081071447", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.228", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a757887b3e1e4c69bea03f4924fc813e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c883fb98-d172-4510-8cf4-07aafdf771af", "external-id": "nsx-vlan-transportzone-570", "segmentation_id": 570, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfd004411-8e", "ovs_interfaceid": "fd004411-8e83-4378-b85c-098463c3e810", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1139.483721] env[62965]: DEBUG oslo_concurrency.lockutils [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Releasing lock "refresh_cache-385648d7-20e3-442c-b6a1-3d972d266487" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1139.720661] env[62965]: DEBUG oslo_concurrency.lockutils [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Releasing lock "refresh_cache-e8ff0b40-b329-404f-a0c8-f123c1fd6d40" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1139.721025] env[62965]: DEBUG nova.compute.manager [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Instance network_info: |[{"id": "1f01a2f6-4965-4721-a5b4-7d1c4b1fde60", "address": "fa:16:3e:51:d0:61", "network": {"id": "71361615-7a84-4c75-9481-afed7dc299d8", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2081071447", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.241", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a757887b3e1e4c69bea03f4924fc813e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c883fb98-d172-4510-8cf4-07aafdf771af", "external-id": "nsx-vlan-transportzone-570", "segmentation_id": 570, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f01a2f6-49", "ovs_interfaceid": "1f01a2f6-4965-4721-a5b4-7d1c4b1fde60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "84a65364-a7f3-4ed6-8496-44a1e968fe12", "address": "fa:16:3e:03:0e:62", "network": {"id": "d61ce9c6-768b-4f8b-8658-00e3b55893eb", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-949610621", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.63", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "a757887b3e1e4c69bea03f4924fc813e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37e8d2ee-abfc-42e2-a8fa-ee5447f1f1da", "external-id": "nsx-vlan-transportzone-813", "segmentation_id": 813, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap84a65364-a7", "ovs_interfaceid": "84a65364-a7f3-4ed6-8496-44a1e968fe12", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "fd004411-8e83-4378-b85c-098463c3e810", "address": "fa:16:3e:1e:0e:d2", "network": {"id": "71361615-7a84-4c75-9481-afed7dc299d8", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2081071447", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.228", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a757887b3e1e4c69bea03f4924fc813e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c883fb98-d172-4510-8cf4-07aafdf771af", "external-id": "nsx-vlan-transportzone-570", "segmentation_id": 570, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfd004411-8e", "ovs_interfaceid": "fd004411-8e83-4378-b85c-098463c3e810", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1139.721353] env[62965]: DEBUG oslo_concurrency.lockutils [req-9705e850-51dd-400c-b747-4de901929061 req-e8fcc841-ce39-4aa0-962a-24f34549f388 service nova] Acquired lock "refresh_cache-e8ff0b40-b329-404f-a0c8-f123c1fd6d40" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1139.721543] env[62965]: DEBUG nova.network.neutron [req-9705e850-51dd-400c-b747-4de901929061 req-e8fcc841-ce39-4aa0-962a-24f34549f388 service nova] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Refreshing network info cache for port fd004411-8e83-4378-b85c-098463c3e810 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1139.722816] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:51:d0:61', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c883fb98-d172-4510-8cf4-07aafdf771af', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1f01a2f6-4965-4721-a5b4-7d1c4b1fde60', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:03:0e:62', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '37e8d2ee-abfc-42e2-a8fa-ee5447f1f1da', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '84a65364-a7f3-4ed6-8496-44a1e968fe12', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:1e:0e:d2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c883fb98-d172-4510-8cf4-07aafdf771af', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fd004411-8e83-4378-b85c-098463c3e810', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1139.733594] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Creating folder: Project (a757887b3e1e4c69bea03f4924fc813e). Parent ref: group-v294931. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1139.736769] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-87c8a092-1afe-493f-8580-db0ea6f6da38 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.747780] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Created folder: Project (a757887b3e1e4c69bea03f4924fc813e) in parent group-v294931. [ 1139.747955] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Creating folder: Instances. Parent ref: group-v295068. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1139.748190] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4c8d7e84-806f-40a8-b280-d09d06999058 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.756799] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Created folder: Instances in parent group-v295068. [ 1139.757028] env[62965]: DEBUG oslo.service.loopingcall [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1139.757214] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1139.757398] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9ea78c3a-272b-439e-bcfd-69774e8d08b2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.782792] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1139.782792] env[62965]: value = "task-1390439" [ 1139.782792] env[62965]: _type = "Task" [ 1139.782792] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.789666] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390439, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.935044] env[62965]: DEBUG nova.network.neutron [req-9705e850-51dd-400c-b747-4de901929061 req-e8fcc841-ce39-4aa0-962a-24f34549f388 service nova] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Updated VIF entry in instance network info cache for port fd004411-8e83-4378-b85c-098463c3e810. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1139.935528] env[62965]: DEBUG nova.network.neutron [req-9705e850-51dd-400c-b747-4de901929061 req-e8fcc841-ce39-4aa0-962a-24f34549f388 service nova] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Updating instance_info_cache with network_info: [{"id": "1f01a2f6-4965-4721-a5b4-7d1c4b1fde60", "address": "fa:16:3e:51:d0:61", "network": {"id": "71361615-7a84-4c75-9481-afed7dc299d8", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2081071447", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.241", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a757887b3e1e4c69bea03f4924fc813e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c883fb98-d172-4510-8cf4-07aafdf771af", "external-id": "nsx-vlan-transportzone-570", "segmentation_id": 570, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f01a2f6-49", "ovs_interfaceid": "1f01a2f6-4965-4721-a5b4-7d1c4b1fde60", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "84a65364-a7f3-4ed6-8496-44a1e968fe12", "address": "fa:16:3e:03:0e:62", "network": {"id": "d61ce9c6-768b-4f8b-8658-00e3b55893eb", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-949610621", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.63", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "a757887b3e1e4c69bea03f4924fc813e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37e8d2ee-abfc-42e2-a8fa-ee5447f1f1da", "external-id": "nsx-vlan-transportzone-813", "segmentation_id": 813, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap84a65364-a7", "ovs_interfaceid": "84a65364-a7f3-4ed6-8496-44a1e968fe12", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "fd004411-8e83-4378-b85c-098463c3e810", "address": "fa:16:3e:1e:0e:d2", "network": {"id": "71361615-7a84-4c75-9481-afed7dc299d8", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2081071447", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.228", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a757887b3e1e4c69bea03f4924fc813e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c883fb98-d172-4510-8cf4-07aafdf771af", "external-id": "nsx-vlan-transportzone-570", "segmentation_id": 570, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfd004411-8e", "ovs_interfaceid": "fd004411-8e83-4378-b85c-098463c3e810", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1140.009055] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb97c8b4-eea6-46c7-94e3-63e201525479 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.030699] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2fb5b14-495e-4f47-bb31-4b19546052ab {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.037803] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Updating instance '385648d7-20e3-442c-b6a1-3d972d266487' progress to 83 {{(pid=62965) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1140.292681] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390439, 'name': CreateVM_Task} progress is 99%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.438278] env[62965]: DEBUG oslo_concurrency.lockutils [req-9705e850-51dd-400c-b747-4de901929061 req-e8fcc841-ce39-4aa0-962a-24f34549f388 service nova] Releasing lock "refresh_cache-e8ff0b40-b329-404f-a0c8-f123c1fd6d40" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1140.544617] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1140.544954] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aa7a9902-4905-47c4-8ca3-f2a48362afa6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.551687] env[62965]: DEBUG oslo_vmware.api [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1140.551687] env[62965]: value = "task-1390440" [ 1140.551687] env[62965]: _type = "Task" [ 1140.551687] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.559343] env[62965]: DEBUG oslo_vmware.api [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390440, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.792985] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390439, 'name': CreateVM_Task} progress is 99%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.061649] env[62965]: DEBUG oslo_vmware.api [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390440, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.294268] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390439, 'name': CreateVM_Task, 'duration_secs': 1.418496} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.294448] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1141.295364] env[62965]: DEBUG oslo_concurrency.lockutils [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1141.295541] env[62965]: DEBUG oslo_concurrency.lockutils [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1141.295890] env[62965]: DEBUG oslo_concurrency.lockutils [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1141.296174] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57ab42c3-e376-442d-b0a0-459f2e0a2970 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.300683] env[62965]: DEBUG oslo_vmware.api [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Waiting for the task: (returnval){ [ 1141.300683] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52789cc9-1d6c-d748-6af1-a2aa641066b3" [ 1141.300683] env[62965]: _type = "Task" [ 1141.300683] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.308736] env[62965]: DEBUG oslo_vmware.api [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52789cc9-1d6c-d748-6af1-a2aa641066b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.364515] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._sync_power_states {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1141.562151] env[62965]: DEBUG oslo_vmware.api [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390440, 'name': PowerOnVM_Task, 'duration_secs': 0.866918} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.562494] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1141.562698] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-9a21bb1e-9f6e-4cda-bc07-40a0483ae0fb tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Updating instance '385648d7-20e3-442c-b6a1-3d972d266487' progress to 100 {{(pid=62965) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1141.811270] env[62965]: DEBUG oslo_vmware.api [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52789cc9-1d6c-d748-6af1-a2aa641066b3, 'name': SearchDatastore_Task, 'duration_secs': 0.010428} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.811402] env[62965]: DEBUG oslo_concurrency.lockutils [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1141.811596] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1141.811838] env[62965]: DEBUG oslo_concurrency.lockutils [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1141.811987] env[62965]: DEBUG oslo_concurrency.lockutils [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1141.812183] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1141.812450] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5f37d627-20bb-4eb6-afca-4dfc19a3d60c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.837021] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1141.837217] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1141.837942] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef23a339-3782-4f3f-9377-2930216e46d5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.843072] env[62965]: DEBUG oslo_vmware.api [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Waiting for the task: (returnval){ [ 1141.843072] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52f5b265-7923-9c4c-0b3b-238d7648476e" [ 1141.843072] env[62965]: _type = "Task" [ 1141.843072] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.850765] env[62965]: DEBUG oslo_vmware.api [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52f5b265-7923-9c4c-0b3b-238d7648476e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.868368] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Getting list of instances from cluster (obj){ [ 1141.868368] env[62965]: value = "domain-c8" [ 1141.868368] env[62965]: _type = "ClusterComputeResource" [ 1141.868368] env[62965]: } {{(pid=62965) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1141.869232] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbdce9ec-624f-4296-a0f9-76c6fe6cad95 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.880393] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Got total of 3 instances {{(pid=62965) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1141.880550] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Triggering sync for uuid 73a7ca0e-b1d3-4464-969b-5111d2506a8b {{(pid=62965) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10667}} [ 1141.880735] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Triggering sync for uuid 385648d7-20e3-442c-b6a1-3d972d266487 {{(pid=62965) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10667}} [ 1141.880888] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Triggering sync for uuid e8ff0b40-b329-404f-a0c8-f123c1fd6d40 {{(pid=62965) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10667}} [ 1141.881186] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Acquiring lock "73a7ca0e-b1d3-4464-969b-5111d2506a8b" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1141.881399] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "73a7ca0e-b1d3-4464-969b-5111d2506a8b" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1141.881648] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Acquiring lock "385648d7-20e3-442c-b6a1-3d972d266487" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1141.881837] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "385648d7-20e3-442c-b6a1-3d972d266487" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1141.882008] env[62965]: INFO nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] During sync_power_state the instance has a pending task (resize_finish). Skip. [ 1141.882181] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "385648d7-20e3-442c-b6a1-3d972d266487" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1141.882371] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Acquiring lock "e8ff0b40-b329-404f-a0c8-f123c1fd6d40" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1141.883162] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dc2789a-f070-4869-9016-b236d8ebaaa5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.352860] env[62965]: DEBUG oslo_vmware.api [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52f5b265-7923-9c4c-0b3b-238d7648476e, 'name': SearchDatastore_Task, 'duration_secs': 0.00891} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.353654] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f73b7821-f7fe-4f07-941f-3a61b84e0d4f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.358483] env[62965]: DEBUG oslo_vmware.api [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Waiting for the task: (returnval){ [ 1142.358483] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]522168f4-46ee-cf10-1519-5e904c9cd9ad" [ 1142.358483] env[62965]: _type = "Task" [ 1142.358483] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.365794] env[62965]: DEBUG oslo_vmware.api [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]522168f4-46ee-cf10-1519-5e904c9cd9ad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.390431] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "73a7ca0e-b1d3-4464-969b-5111d2506a8b" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.509s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.868874] env[62965]: DEBUG oslo_vmware.api [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]522168f4-46ee-cf10-1519-5e904c9cd9ad, 'name': SearchDatastore_Task, 'duration_secs': 0.009199} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.869239] env[62965]: DEBUG oslo_concurrency.lockutils [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1142.869363] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] e8ff0b40-b329-404f-a0c8-f123c1fd6d40/e8ff0b40-b329-404f-a0c8-f123c1fd6d40.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1142.869626] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-345cf2e6-d816-406e-b8b4-74ee11e3bb3b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.877090] env[62965]: DEBUG oslo_vmware.api [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Waiting for the task: (returnval){ [ 1142.877090] env[62965]: value = "task-1390441" [ 1142.877090] env[62965]: _type = "Task" [ 1142.877090] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.884156] env[62965]: DEBUG oslo_vmware.api [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': task-1390441, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.386803] env[62965]: DEBUG oslo_vmware.api [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': task-1390441, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.887365] env[62965]: DEBUG oslo_vmware.api [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': task-1390441, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.533363} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.887718] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] e8ff0b40-b329-404f-a0c8-f123c1fd6d40/e8ff0b40-b329-404f-a0c8-f123c1fd6d40.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1143.887869] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1143.888105] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-44da3271-56be-473f-8c33-438473251b89 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.896331] env[62965]: DEBUG oslo_vmware.api [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Waiting for the task: (returnval){ [ 1143.896331] env[62965]: value = "task-1390442" [ 1143.896331] env[62965]: _type = "Task" [ 1143.896331] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.903691] env[62965]: DEBUG oslo_vmware.api [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': task-1390442, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.406084] env[62965]: DEBUG oslo_vmware.api [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': task-1390442, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065642} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.406371] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1144.407159] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aec64b6a-20f6-4e31-98bd-ba6d7ec8f980 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.437259] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] e8ff0b40-b329-404f-a0c8-f123c1fd6d40/e8ff0b40-b329-404f-a0c8-f123c1fd6d40.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1144.437579] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6a0f5b1c-384b-4a2f-8521-726ace98d2ff {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.460737] env[62965]: DEBUG oslo_vmware.api [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Waiting for the task: (returnval){ [ 1144.460737] env[62965]: value = "task-1390443" [ 1144.460737] env[62965]: _type = "Task" [ 1144.460737] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.468597] env[62965]: DEBUG oslo_vmware.api [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': task-1390443, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.488297] env[62965]: DEBUG nova.network.neutron [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Port 7de6b63c-f2c3-4a0b-b399-73075c5e09a5 binding to destination host cpu-1 is already ACTIVE {{(pid=62965) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1144.488562] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "refresh_cache-385648d7-20e3-442c-b6a1-3d972d266487" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1144.488732] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquired lock "refresh_cache-385648d7-20e3-442c-b6a1-3d972d266487" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1144.488889] env[62965]: DEBUG nova.network.neutron [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1144.971383] env[62965]: DEBUG oslo_vmware.api [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': task-1390443, 'name': ReconfigVM_Task, 'duration_secs': 0.404323} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.971704] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Reconfigured VM instance instance-0000006c to attach disk [datastore1] e8ff0b40-b329-404f-a0c8-f123c1fd6d40/e8ff0b40-b329-404f-a0c8-f123c1fd6d40.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1144.972335] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7a614067-6d83-4817-9f0e-c2e784893760 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.978694] env[62965]: DEBUG oslo_vmware.api [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Waiting for the task: (returnval){ [ 1144.978694] env[62965]: value = "task-1390444" [ 1144.978694] env[62965]: _type = "Task" [ 1144.978694] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.986287] env[62965]: DEBUG oslo_vmware.api [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': task-1390444, 'name': Rename_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.219053] env[62965]: DEBUG nova.network.neutron [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Updating instance_info_cache with network_info: [{"id": "7de6b63c-f2c3-4a0b-b399-73075c5e09a5", "address": "fa:16:3e:6c:49:73", "network": {"id": "4649fed7-c2ca-43a9-a183-7c99423da726", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1978698173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fab418e2fe1420793517663574b43bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43ad01d2-c7dd-453c-a929-8ad76294d13c", "external-id": "nsx-vlan-transportzone-176", "segmentation_id": 176, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7de6b63c-f2", "ovs_interfaceid": "7de6b63c-f2c3-4a0b-b399-73075c5e09a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1145.489236] env[62965]: DEBUG oslo_vmware.api [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': task-1390444, 'name': Rename_Task, 'duration_secs': 0.174353} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.489485] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1145.489714] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0ae71fe2-6e88-4337-b956-10f5b9585f6c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.496058] env[62965]: DEBUG oslo_vmware.api [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Waiting for the task: (returnval){ [ 1145.496058] env[62965]: value = "task-1390445" [ 1145.496058] env[62965]: _type = "Task" [ 1145.496058] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.503213] env[62965]: DEBUG oslo_vmware.api [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': task-1390445, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.722394] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Releasing lock "refresh_cache-385648d7-20e3-442c-b6a1-3d972d266487" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1146.005763] env[62965]: DEBUG oslo_vmware.api [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': task-1390445, 'name': PowerOnVM_Task, 'duration_secs': 0.474429} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.006101] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1146.006296] env[62965]: INFO nova.compute.manager [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Took 13.58 seconds to spawn the instance on the hypervisor. [ 1146.006490] env[62965]: DEBUG nova.compute.manager [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1146.007248] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97da0295-4b0d-4536-b1a8-f65d9525ced7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.226535] env[62965]: DEBUG nova.compute.manager [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=62965) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:900}} [ 1146.226741] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1146.226956] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1146.526409] env[62965]: INFO nova.compute.manager [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Took 18.30 seconds to build instance. [ 1146.729594] env[62965]: DEBUG nova.objects.instance [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lazy-loading 'migration_context' on Instance uuid 385648d7-20e3-442c-b6a1-3d972d266487 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1147.028845] env[62965]: DEBUG oslo_concurrency.lockutils [None req-858debcb-3690-4af6-a794-6a5a9e1f9076 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Lock "e8ff0b40-b329-404f-a0c8-f123c1fd6d40" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.813s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1147.029194] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "e8ff0b40-b329-404f-a0c8-f123c1fd6d40" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 5.147s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1147.029380] env[62965]: INFO nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] During sync_power_state the instance has a pending task (spawning). Skip. [ 1147.029569] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "e8ff0b40-b329-404f-a0c8-f123c1fd6d40" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1147.295171] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adcc6537-ffa3-40a9-841b-965a3b64ed33 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.302718] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c14bc242-ef51-44b8-b140-799fe4e35c8c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.331897] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f9c5c6c-e635-4124-b7d9-632343a7492f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.338919] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1749ac41-fe3e-4117-baa6-7f265c332408 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.354761] env[62965]: DEBUG nova.compute.provider_tree [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1147.622606] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1147.622831] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Starting heal instance info cache {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10257}} [ 1147.623056] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Rebuilding the list of instances to heal {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10261}} [ 1147.715102] env[62965]: DEBUG oslo_concurrency.lockutils [None req-61d4f5e1-a462-49e4-88ce-13252a2671fa tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Acquiring lock "e8ff0b40-b329-404f-a0c8-f123c1fd6d40" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1147.715372] env[62965]: DEBUG oslo_concurrency.lockutils [None req-61d4f5e1-a462-49e4-88ce-13252a2671fa tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Lock "e8ff0b40-b329-404f-a0c8-f123c1fd6d40" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1147.715588] env[62965]: DEBUG oslo_concurrency.lockutils [None req-61d4f5e1-a462-49e4-88ce-13252a2671fa tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Acquiring lock "e8ff0b40-b329-404f-a0c8-f123c1fd6d40-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1147.715778] env[62965]: DEBUG oslo_concurrency.lockutils [None req-61d4f5e1-a462-49e4-88ce-13252a2671fa tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Lock "e8ff0b40-b329-404f-a0c8-f123c1fd6d40-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1147.715957] env[62965]: DEBUG oslo_concurrency.lockutils [None req-61d4f5e1-a462-49e4-88ce-13252a2671fa tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Lock "e8ff0b40-b329-404f-a0c8-f123c1fd6d40-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1147.718101] env[62965]: INFO nova.compute.manager [None req-61d4f5e1-a462-49e4-88ce-13252a2671fa tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Terminating instance [ 1147.858007] env[62965]: DEBUG nova.scheduler.client.report [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1148.127122] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Skipping network cache update for instance because it is being deleted. {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10274}} [ 1148.153642] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Acquiring lock "refresh_cache-73a7ca0e-b1d3-4464-969b-5111d2506a8b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1148.153850] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Acquired lock "refresh_cache-73a7ca0e-b1d3-4464-969b-5111d2506a8b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1148.154050] env[62965]: DEBUG nova.network.neutron [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Forcefully refreshing network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1148.154226] env[62965]: DEBUG nova.objects.instance [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lazy-loading 'info_cache' on Instance uuid 73a7ca0e-b1d3-4464-969b-5111d2506a8b {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1148.221388] env[62965]: DEBUG nova.compute.manager [None req-61d4f5e1-a462-49e4-88ce-13252a2671fa tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1148.221605] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-61d4f5e1-a462-49e4-88ce-13252a2671fa tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1148.222576] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d47f077b-a679-489e-810c-3a90cce7c511 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.229940] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-61d4f5e1-a462-49e4-88ce-13252a2671fa tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1148.230199] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ffeae654-e48a-4bd0-89c3-76a04dfbf0c3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.236367] env[62965]: DEBUG oslo_vmware.api [None req-61d4f5e1-a462-49e4-88ce-13252a2671fa tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Waiting for the task: (returnval){ [ 1148.236367] env[62965]: value = "task-1390446" [ 1148.236367] env[62965]: _type = "Task" [ 1148.236367] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.244045] env[62965]: DEBUG oslo_vmware.api [None req-61d4f5e1-a462-49e4-88ce-13252a2671fa tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': task-1390446, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.746156] env[62965]: DEBUG oslo_vmware.api [None req-61d4f5e1-a462-49e4-88ce-13252a2671fa tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': task-1390446, 'name': PowerOffVM_Task, 'duration_secs': 0.190953} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.746374] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-61d4f5e1-a462-49e4-88ce-13252a2671fa tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1148.746544] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-61d4f5e1-a462-49e4-88ce-13252a2671fa tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1148.746827] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cf1b8ec0-c3b4-4f3a-8606-4d04582d0e40 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.871083] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.644s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1149.089389] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-61d4f5e1-a462-49e4-88ce-13252a2671fa tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1149.089630] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-61d4f5e1-a462-49e4-88ce-13252a2671fa tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Deleting contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1149.089897] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-61d4f5e1-a462-49e4-88ce-13252a2671fa tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Deleting the datastore file [datastore1] e8ff0b40-b329-404f-a0c8-f123c1fd6d40 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1149.090156] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cc376e06-eea1-4b58-bca8-811a4585c543 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.097397] env[62965]: DEBUG oslo_vmware.api [None req-61d4f5e1-a462-49e4-88ce-13252a2671fa tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Waiting for the task: (returnval){ [ 1149.097397] env[62965]: value = "task-1390448" [ 1149.097397] env[62965]: _type = "Task" [ 1149.097397] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.105444] env[62965]: DEBUG oslo_vmware.api [None req-61d4f5e1-a462-49e4-88ce-13252a2671fa tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': task-1390448, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.607903] env[62965]: DEBUG oslo_vmware.api [None req-61d4f5e1-a462-49e4-88ce-13252a2671fa tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': task-1390448, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.17641} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.608409] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-61d4f5e1-a462-49e4-88ce-13252a2671fa tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1149.608409] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-61d4f5e1-a462-49e4-88ce-13252a2671fa tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Deleted contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1149.608589] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-61d4f5e1-a462-49e4-88ce-13252a2671fa tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1149.608749] env[62965]: INFO nova.compute.manager [None req-61d4f5e1-a462-49e4-88ce-13252a2671fa tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Took 1.39 seconds to destroy the instance on the hypervisor. [ 1149.608975] env[62965]: DEBUG oslo.service.loopingcall [None req-61d4f5e1-a462-49e4-88ce-13252a2671fa tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1149.609186] env[62965]: DEBUG nova.compute.manager [-] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1149.609293] env[62965]: DEBUG nova.network.neutron [-] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1149.846917] env[62965]: DEBUG nova.compute.manager [req-901b16cd-a01d-40de-97e5-1e6407b3d18f req-f5c9e6c5-5014-4115-915e-eed9634ad941 service nova] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Received event network-vif-deleted-1f01a2f6-4965-4721-a5b4-7d1c4b1fde60 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1149.847259] env[62965]: INFO nova.compute.manager [req-901b16cd-a01d-40de-97e5-1e6407b3d18f req-f5c9e6c5-5014-4115-915e-eed9634ad941 service nova] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Neutron deleted interface 1f01a2f6-4965-4721-a5b4-7d1c4b1fde60; detaching it from the instance and deleting it from the info cache [ 1149.847727] env[62965]: DEBUG nova.network.neutron [req-901b16cd-a01d-40de-97e5-1e6407b3d18f req-f5c9e6c5-5014-4115-915e-eed9634ad941 service nova] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Updating instance_info_cache with network_info: [{"id": "84a65364-a7f3-4ed6-8496-44a1e968fe12", "address": "fa:16:3e:03:0e:62", "network": {"id": "d61ce9c6-768b-4f8b-8658-00e3b55893eb", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-949610621", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.63", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "a757887b3e1e4c69bea03f4924fc813e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37e8d2ee-abfc-42e2-a8fa-ee5447f1f1da", "external-id": "nsx-vlan-transportzone-813", "segmentation_id": 813, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap84a65364-a7", "ovs_interfaceid": "84a65364-a7f3-4ed6-8496-44a1e968fe12", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "fd004411-8e83-4378-b85c-098463c3e810", "address": "fa:16:3e:1e:0e:d2", "network": {"id": "71361615-7a84-4c75-9481-afed7dc299d8", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2081071447", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.228", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a757887b3e1e4c69bea03f4924fc813e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c883fb98-d172-4510-8cf4-07aafdf771af", "external-id": "nsx-vlan-transportzone-570", "segmentation_id": 570, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfd004411-8e", "ovs_interfaceid": "fd004411-8e83-4378-b85c-098463c3e810", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1149.884136] env[62965]: DEBUG nova.network.neutron [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Updating instance_info_cache with network_info: [{"id": "5f939a36-bc5f-4fd7-b29e-011f72b3256c", "address": "fa:16:3e:e3:06:99", "network": {"id": "49dd42b0-6397-4fdd-a319-ed920c72a5af", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-373172566-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.200", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d91c99b14018469e8998fd1ccc0fab4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f77ff7a1-209c-4f3f-b2a0-fd817741e739", "external-id": "nsx-vlan-transportzone-935", "segmentation_id": 935, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f939a36-bc", "ovs_interfaceid": "5f939a36-bc5f-4fd7-b29e-011f72b3256c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1150.351455] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dd4d255d-5031-4d63-a6b6-f0a7f4c0f177 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.361118] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffca5a1d-7328-4d9c-aa76-4b8dd111f095 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.385132] env[62965]: DEBUG nova.compute.manager [req-901b16cd-a01d-40de-97e5-1e6407b3d18f req-f5c9e6c5-5014-4115-915e-eed9634ad941 service nova] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Detach interface failed, port_id=1f01a2f6-4965-4721-a5b4-7d1c4b1fde60, reason: Instance e8ff0b40-b329-404f-a0c8-f123c1fd6d40 could not be found. {{(pid=62965) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11294}} [ 1150.386723] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Releasing lock "refresh_cache-73a7ca0e-b1d3-4464-969b-5111d2506a8b" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1150.386908] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Updated the network info_cache for instance {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10328}} [ 1150.406888] env[62965]: INFO nova.compute.manager [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Swapping old allocation on dict_keys(['20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8']) held by migration ef5466ea-156c-429a-ae78-424d5f9443a8 for instance [ 1150.429248] env[62965]: DEBUG nova.scheduler.client.report [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Overwriting current allocation {'allocations': {'20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8': {'resources': {'VCPU': 1, 'MEMORY_MB': 256, 'DISK_GB': 1}, 'generation': 143}}, 'project_id': '8fab418e2fe1420793517663574b43bb', 'user_id': 'e9d9f79c936143f5815fcb99755bbd4e', 'consumer_generation': 1} on consumer 385648d7-20e3-442c-b6a1-3d972d266487 {{(pid=62965) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2033}} [ 1150.510237] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "refresh_cache-385648d7-20e3-442c-b6a1-3d972d266487" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1150.510429] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquired lock "refresh_cache-385648d7-20e3-442c-b6a1-3d972d266487" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1150.510627] env[62965]: DEBUG nova.network.neutron [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1150.643395] env[62965]: DEBUG nova.network.neutron [-] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1151.104193] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1151.104429] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1151.147956] env[62965]: INFO nova.compute.manager [-] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Took 1.54 seconds to deallocate network for instance. [ 1151.231366] env[62965]: DEBUG nova.network.neutron [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Updating instance_info_cache with network_info: [{"id": "7de6b63c-f2c3-4a0b-b399-73075c5e09a5", "address": "fa:16:3e:6c:49:73", "network": {"id": "4649fed7-c2ca-43a9-a183-7c99423da726", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1978698173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fab418e2fe1420793517663574b43bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43ad01d2-c7dd-453c-a929-8ad76294d13c", "external-id": "nsx-vlan-transportzone-176", "segmentation_id": 176, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7de6b63c-f2", "ovs_interfaceid": "7de6b63c-f2c3-4a0b-b399-73075c5e09a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1151.654325] env[62965]: DEBUG oslo_concurrency.lockutils [None req-61d4f5e1-a462-49e4-88ce-13252a2671fa tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1151.654603] env[62965]: DEBUG oslo_concurrency.lockutils [None req-61d4f5e1-a462-49e4-88ce-13252a2671fa tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1151.654838] env[62965]: DEBUG nova.objects.instance [None req-61d4f5e1-a462-49e4-88ce-13252a2671fa tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Lazy-loading 'resources' on Instance uuid e8ff0b40-b329-404f-a0c8-f123c1fd6d40 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1151.733583] env[62965]: DEBUG oslo_concurrency.lockutils [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Releasing lock "refresh_cache-385648d7-20e3-442c-b6a1-3d972d266487" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1151.734105] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1151.734394] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-548ab507-5428-4bf4-8b06-4452e1b05426 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.742330] env[62965]: DEBUG oslo_vmware.api [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1151.742330] env[62965]: value = "task-1390449" [ 1151.742330] env[62965]: _type = "Task" [ 1151.742330] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.750683] env[62965]: DEBUG oslo_vmware.api [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390449, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.866944] env[62965]: DEBUG nova.compute.manager [req-eb0f37a9-d156-4bd1-be2f-6747a1798e23 req-de3bc349-fbf6-43be-961c-eee38e1a7d07 service nova] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Received event network-vif-deleted-84a65364-a7f3-4ed6-8496-44a1e968fe12 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1151.867321] env[62965]: DEBUG nova.compute.manager [req-eb0f37a9-d156-4bd1-be2f-6747a1798e23 req-de3bc349-fbf6-43be-961c-eee38e1a7d07 service nova] [instance: e8ff0b40-b329-404f-a0c8-f123c1fd6d40] Received event network-vif-deleted-fd004411-8e83-4378-b85c-098463c3e810 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1152.105085] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1152.105085] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62965) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10876}} [ 1152.208333] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f0b7ed7-61c8-4d48-a227-6917fce75450 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.215787] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d10d700-0b0c-4b6a-8fc9-0901a6f898f6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.249208] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7adfb84b-8ad2-4f67-bb53-8fb8c28ec081 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.256676] env[62965]: DEBUG oslo_vmware.api [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390449, 'name': PowerOffVM_Task, 'duration_secs': 0.205891} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.258555] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1152.259200] env[62965]: DEBUG nova.virt.hardware [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1152.259538] env[62965]: DEBUG nova.virt.hardware [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1152.259538] env[62965]: DEBUG nova.virt.hardware [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1152.259728] env[62965]: DEBUG nova.virt.hardware [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1152.259874] env[62965]: DEBUG nova.virt.hardware [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1152.260031] env[62965]: DEBUG nova.virt.hardware [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1152.260243] env[62965]: DEBUG nova.virt.hardware [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1152.260471] env[62965]: DEBUG nova.virt.hardware [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1152.260657] env[62965]: DEBUG nova.virt.hardware [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1152.260816] env[62965]: DEBUG nova.virt.hardware [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1152.260989] env[62965]: DEBUG nova.virt.hardware [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1152.265832] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-08fe1684-8902-4b81-aed9-57581ac6160a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.276459] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84533c90-cd8b-4789-a593-e6eb58b009ea {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.290373] env[62965]: DEBUG nova.compute.provider_tree [None req-61d4f5e1-a462-49e4-88ce-13252a2671fa tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1152.292703] env[62965]: DEBUG oslo_vmware.api [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1152.292703] env[62965]: value = "task-1390450" [ 1152.292703] env[62965]: _type = "Task" [ 1152.292703] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.300388] env[62965]: DEBUG oslo_vmware.api [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390450, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.794186] env[62965]: DEBUG nova.scheduler.client.report [None req-61d4f5e1-a462-49e4-88ce-13252a2671fa tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1152.806610] env[62965]: DEBUG oslo_vmware.api [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390450, 'name': ReconfigVM_Task, 'duration_secs': 0.135187} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.807448] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d16f07a-61a4-4c68-b73d-d56969b1419c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.826158] env[62965]: DEBUG nova.virt.hardware [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1152.826402] env[62965]: DEBUG nova.virt.hardware [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1152.826561] env[62965]: DEBUG nova.virt.hardware [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1152.826746] env[62965]: DEBUG nova.virt.hardware [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1152.826894] env[62965]: DEBUG nova.virt.hardware [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1152.827058] env[62965]: DEBUG nova.virt.hardware [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1152.827280] env[62965]: DEBUG nova.virt.hardware [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1152.827441] env[62965]: DEBUG nova.virt.hardware [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1152.827611] env[62965]: DEBUG nova.virt.hardware [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1152.827776] env[62965]: DEBUG nova.virt.hardware [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1152.827952] env[62965]: DEBUG nova.virt.hardware [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1152.828726] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-176966a2-1ec4-429b-a8bb-3d050054ba7b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.834282] env[62965]: DEBUG oslo_vmware.api [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1152.834282] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52748a85-1b58-0800-63f4-23dd8e618a69" [ 1152.834282] env[62965]: _type = "Task" [ 1152.834282] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.843900] env[62965]: DEBUG oslo_vmware.api [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52748a85-1b58-0800-63f4-23dd8e618a69, 'name': SearchDatastore_Task, 'duration_secs': 0.007619} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.849210] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Reconfiguring VM instance instance-0000006a to detach disk 2000 {{(pid=62965) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1152.849472] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-605bc302-4161-4b63-93c0-315192335f0c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.867065] env[62965]: DEBUG oslo_vmware.api [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1152.867065] env[62965]: value = "task-1390451" [ 1152.867065] env[62965]: _type = "Task" [ 1152.867065] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.874501] env[62965]: DEBUG oslo_vmware.api [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390451, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.105119] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1153.105389] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1153.303073] env[62965]: DEBUG oslo_concurrency.lockutils [None req-61d4f5e1-a462-49e4-88ce-13252a2671fa tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.648s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1153.321053] env[62965]: INFO nova.scheduler.client.report [None req-61d4f5e1-a462-49e4-88ce-13252a2671fa tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Deleted allocations for instance e8ff0b40-b329-404f-a0c8-f123c1fd6d40 [ 1153.376967] env[62965]: DEBUG oslo_vmware.api [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390451, 'name': ReconfigVM_Task, 'duration_secs': 0.181393} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.377264] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Reconfigured VM instance instance-0000006a to detach disk 2000 {{(pid=62965) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1153.378022] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ec17185-2a62-48cc-a059-be445b9dc02e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.399508] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] 385648d7-20e3-442c-b6a1-3d972d266487/385648d7-20e3-442c-b6a1-3d972d266487.vmdk or device None with type thin {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1153.399964] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cbeb6abd-316d-48b8-b0c5-9aa33a990d6d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.417888] env[62965]: DEBUG oslo_vmware.api [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1153.417888] env[62965]: value = "task-1390452" [ 1153.417888] env[62965]: _type = "Task" [ 1153.417888] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.425230] env[62965]: DEBUG oslo_vmware.api [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390452, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.828418] env[62965]: DEBUG oslo_concurrency.lockutils [None req-61d4f5e1-a462-49e4-88ce-13252a2671fa tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Lock "e8ff0b40-b329-404f-a0c8-f123c1fd6d40" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.113s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1153.927898] env[62965]: DEBUG oslo_vmware.api [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390452, 'name': ReconfigVM_Task, 'duration_secs': 0.24727} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.928400] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Reconfigured VM instance instance-0000006a to attach disk [datastore1] 385648d7-20e3-442c-b6a1-3d972d266487/385648d7-20e3-442c-b6a1-3d972d266487.vmdk or device None with type thin {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1153.929390] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2162801d-d7d5-45d9-8a63-6051eab02e4d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.949708] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db513279-e11c-4c9c-98b0-47f418a62b2f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.967789] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ecf67a2-00e0-4d84-bed5-51703d2b1606 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.987648] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b8be7dd-7ddb-4ae5-96a0-2931e10d5712 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.994408] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1153.994723] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cbcad98e-2636-4e77-81fb-90bca966f094 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.001117] env[62965]: DEBUG oslo_vmware.api [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1154.001117] env[62965]: value = "task-1390453" [ 1154.001117] env[62965]: _type = "Task" [ 1154.001117] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.009897] env[62965]: DEBUG oslo_vmware.api [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390453, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.099233] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1154.511626] env[62965]: DEBUG oslo_vmware.api [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390453, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.013556] env[62965]: DEBUG oslo_vmware.api [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390453, 'name': PowerOnVM_Task, 'duration_secs': 0.564565} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.013873] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1155.104809] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1155.608047] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.608273] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.608440] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.608593] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62965) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1155.609544] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d21e93f-861e-4265-8bc5-3df168cbade4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.617490] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-648487ac-0771-4fdb-8545-60965c9c24c1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.630961] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff42347b-e3ea-4991-ac34-5fab3f79f3b9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.637258] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af2efb89-2c02-4adb-ba5b-586e6a8fc823 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.666710] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181343MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62965) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1155.666885] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.667100] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1156.026991] env[62965]: INFO nova.compute.manager [None req-f58e3ae8-7c78-49ba-ba29-afada8bde4e3 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Updating instance to original state: 'active' [ 1156.030546] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Acquiring lock "79a4df85-7cb0-4ace-ad29-faec14d3a2f8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1156.030546] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Lock "79a4df85-7cb0-4ace-ad29-faec14d3a2f8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1156.535631] env[62965]: DEBUG nova.compute.manager [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1156.692193] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 73a7ca0e-b1d3-4464-969b-5111d2506a8b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1156.692360] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 385648d7-20e3-442c-b6a1-3d972d266487 actively managed on this compute host and has allocations in placement: {'resources': {'VCPU': 1, 'MEMORY_MB': 192, 'DISK_GB': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1156.963354] env[62965]: DEBUG oslo_concurrency.lockutils [None req-49059b49-a8ba-445a-9edd-b4268975065b tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "385648d7-20e3-442c-b6a1-3d972d266487" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1156.963556] env[62965]: DEBUG oslo_concurrency.lockutils [None req-49059b49-a8ba-445a-9edd-b4268975065b tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "385648d7-20e3-442c-b6a1-3d972d266487" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1156.963803] env[62965]: DEBUG oslo_concurrency.lockutils [None req-49059b49-a8ba-445a-9edd-b4268975065b tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "385648d7-20e3-442c-b6a1-3d972d266487-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1156.963997] env[62965]: DEBUG oslo_concurrency.lockutils [None req-49059b49-a8ba-445a-9edd-b4268975065b tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "385648d7-20e3-442c-b6a1-3d972d266487-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1156.964188] env[62965]: DEBUG oslo_concurrency.lockutils [None req-49059b49-a8ba-445a-9edd-b4268975065b tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "385648d7-20e3-442c-b6a1-3d972d266487-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1156.966285] env[62965]: INFO nova.compute.manager [None req-49059b49-a8ba-445a-9edd-b4268975065b tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Terminating instance [ 1157.058069] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1157.195608] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 79a4df85-7cb0-4ace-ad29-faec14d3a2f8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1157.195608] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=62965) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1157.195837] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=62965) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1157.240218] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65bc2658-0c7f-4992-aa22-28e916672ca9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.247380] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b63133c5-e398-42c9-95f2-e0dc71b88d03 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.277336] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1787a797-712a-432b-ac3e-f867c9578b2b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.284242] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b9b6b1a-e3e9-4f19-9237-dc010e8d26fc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.296803] env[62965]: DEBUG nova.compute.provider_tree [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1157.469531] env[62965]: DEBUG nova.compute.manager [None req-49059b49-a8ba-445a-9edd-b4268975065b tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1157.469710] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-49059b49-a8ba-445a-9edd-b4268975065b tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1157.470628] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87961655-7b18-4db3-b7b7-4c4bd07f1cf1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.478370] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-49059b49-a8ba-445a-9edd-b4268975065b tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1157.478598] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-806f5904-8473-44ec-9314-1a1dec9595a0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.485675] env[62965]: DEBUG oslo_vmware.api [None req-49059b49-a8ba-445a-9edd-b4268975065b tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1157.485675] env[62965]: value = "task-1390454" [ 1157.485675] env[62965]: _type = "Task" [ 1157.485675] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.497659] env[62965]: DEBUG oslo_vmware.api [None req-49059b49-a8ba-445a-9edd-b4268975065b tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390454, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.799750] env[62965]: DEBUG nova.scheduler.client.report [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1157.995693] env[62965]: DEBUG oslo_vmware.api [None req-49059b49-a8ba-445a-9edd-b4268975065b tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390454, 'name': PowerOffVM_Task, 'duration_secs': 0.281167} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.995922] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-49059b49-a8ba-445a-9edd-b4268975065b tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1157.996104] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-49059b49-a8ba-445a-9edd-b4268975065b tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1157.996385] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7495822b-e717-4b66-ba98-e7242aad7492 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.304874] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62965) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1158.305198] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.638s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1158.305437] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.247s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1158.306882] env[62965]: INFO nova.compute.claims [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1159.309078] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1159.358234] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90e428c8-f65b-40ab-b48b-1f26f4723d5b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.366031] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba0fd8d5-2301-442c-a952-584647eff450 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.395897] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-518b32a8-02ff-44c5-86cd-5621779ad0a2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.402976] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac840722-4421-4fd4-9da4-63e50bcbb165 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.415622] env[62965]: DEBUG nova.compute.provider_tree [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1159.683090] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-49059b49-a8ba-445a-9edd-b4268975065b tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1159.683332] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-49059b49-a8ba-445a-9edd-b4268975065b tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Deleting contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1159.683519] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-49059b49-a8ba-445a-9edd-b4268975065b tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Deleting the datastore file [datastore1] 385648d7-20e3-442c-b6a1-3d972d266487 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1159.683842] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b730bf05-7fad-4d4c-b2b5-df27c83fd1a8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.691393] env[62965]: DEBUG oslo_vmware.api [None req-49059b49-a8ba-445a-9edd-b4268975065b tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1159.691393] env[62965]: value = "task-1390456" [ 1159.691393] env[62965]: _type = "Task" [ 1159.691393] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.698555] env[62965]: DEBUG oslo_vmware.api [None req-49059b49-a8ba-445a-9edd-b4268975065b tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390456, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.919011] env[62965]: DEBUG nova.scheduler.client.report [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1160.201365] env[62965]: DEBUG oslo_vmware.api [None req-49059b49-a8ba-445a-9edd-b4268975065b tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390456, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.177364} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.201640] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-49059b49-a8ba-445a-9edd-b4268975065b tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1160.201832] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-49059b49-a8ba-445a-9edd-b4268975065b tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Deleted contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1160.202023] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-49059b49-a8ba-445a-9edd-b4268975065b tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1160.202204] env[62965]: INFO nova.compute.manager [None req-49059b49-a8ba-445a-9edd-b4268975065b tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Took 2.73 seconds to destroy the instance on the hypervisor. [ 1160.202443] env[62965]: DEBUG oslo.service.loopingcall [None req-49059b49-a8ba-445a-9edd-b4268975065b tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1160.202636] env[62965]: DEBUG nova.compute.manager [-] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1160.202732] env[62965]: DEBUG nova.network.neutron [-] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1160.424245] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.119s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1160.424785] env[62965]: DEBUG nova.compute.manager [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1160.645477] env[62965]: DEBUG nova.compute.manager [req-384fa74d-d206-4520-8e36-d2e1336ca8d3 req-b04cde7b-82fb-4afb-bd34-fc95fb75643a service nova] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Received event network-vif-deleted-7de6b63c-f2c3-4a0b-b399-73075c5e09a5 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1160.645680] env[62965]: INFO nova.compute.manager [req-384fa74d-d206-4520-8e36-d2e1336ca8d3 req-b04cde7b-82fb-4afb-bd34-fc95fb75643a service nova] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Neutron deleted interface 7de6b63c-f2c3-4a0b-b399-73075c5e09a5; detaching it from the instance and deleting it from the info cache [ 1160.645858] env[62965]: DEBUG nova.network.neutron [req-384fa74d-d206-4520-8e36-d2e1336ca8d3 req-b04cde7b-82fb-4afb-bd34-fc95fb75643a service nova] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1160.929646] env[62965]: DEBUG nova.compute.utils [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1160.931192] env[62965]: DEBUG nova.compute.manager [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1160.931300] env[62965]: DEBUG nova.network.neutron [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1160.981852] env[62965]: DEBUG nova.policy [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '777f044a91c14b5ca3f92b12a9569f94', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a757887b3e1e4c69bea03f4924fc813e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 1161.082873] env[62965]: DEBUG nova.network.neutron [-] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1161.148661] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e935ce87-b9c6-4fb6-8f8d-bfbe1c2f4f27 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.162428] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85adfa64-975b-40b2-b10d-024304d1fa34 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.187698] env[62965]: DEBUG nova.compute.manager [req-384fa74d-d206-4520-8e36-d2e1336ca8d3 req-b04cde7b-82fb-4afb-bd34-fc95fb75643a service nova] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Detach interface failed, port_id=7de6b63c-f2c3-4a0b-b399-73075c5e09a5, reason: Instance 385648d7-20e3-442c-b6a1-3d972d266487 could not be found. {{(pid=62965) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11294}} [ 1161.223285] env[62965]: DEBUG nova.network.neutron [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Successfully created port: c265581c-e8f5-4b97-a87c-3c686b21e331 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1161.435009] env[62965]: DEBUG nova.compute.manager [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1161.448736] env[62965]: DEBUG nova.network.neutron [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Successfully created port: a662d925-10f0-40be-bc51-13a18e973e25 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1161.585108] env[62965]: INFO nova.compute.manager [-] [instance: 385648d7-20e3-442c-b6a1-3d972d266487] Took 1.38 seconds to deallocate network for instance. [ 1162.092349] env[62965]: DEBUG oslo_concurrency.lockutils [None req-49059b49-a8ba-445a-9edd-b4268975065b tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1162.092626] env[62965]: DEBUG oslo_concurrency.lockutils [None req-49059b49-a8ba-445a-9edd-b4268975065b tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1162.092855] env[62965]: DEBUG nova.objects.instance [None req-49059b49-a8ba-445a-9edd-b4268975065b tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lazy-loading 'resources' on Instance uuid 385648d7-20e3-442c-b6a1-3d972d266487 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1162.443523] env[62965]: DEBUG nova.compute.manager [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1162.468634] env[62965]: DEBUG nova.virt.hardware [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1162.468925] env[62965]: DEBUG nova.virt.hardware [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1162.469101] env[62965]: DEBUG nova.virt.hardware [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1162.469296] env[62965]: DEBUG nova.virt.hardware [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1162.469444] env[62965]: DEBUG nova.virt.hardware [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1162.469593] env[62965]: DEBUG nova.virt.hardware [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1162.469800] env[62965]: DEBUG nova.virt.hardware [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1162.469963] env[62965]: DEBUG nova.virt.hardware [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1162.470148] env[62965]: DEBUG nova.virt.hardware [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1162.470315] env[62965]: DEBUG nova.virt.hardware [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1162.470490] env[62965]: DEBUG nova.virt.hardware [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1162.471376] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-396f9500-af56-4d8d-a73c-7fdc43750dcd {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.479583] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a832841-2c70-477f-b2ef-f24cce54538a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.646022] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-917e445d-3f65-404d-8efb-992d44bac92a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.653700] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79b1f8d7-c927-44a9-8a0d-a8d661508a44 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.683844] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff5edb7d-5710-441b-a76c-57f826ebc31e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.690596] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b25d448-9b7f-4d5d-9705-6e369ab6b731 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.703765] env[62965]: DEBUG nova.compute.provider_tree [None req-49059b49-a8ba-445a-9edd-b4268975065b tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1162.809410] env[62965]: DEBUG nova.compute.manager [req-412dbbab-dea3-491d-909d-b5bde3761e09 req-16bb6662-06a2-4a30-a6cf-056c0b84ac0c service nova] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Received event network-vif-plugged-c265581c-e8f5-4b97-a87c-3c686b21e331 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1162.809638] env[62965]: DEBUG oslo_concurrency.lockutils [req-412dbbab-dea3-491d-909d-b5bde3761e09 req-16bb6662-06a2-4a30-a6cf-056c0b84ac0c service nova] Acquiring lock "79a4df85-7cb0-4ace-ad29-faec14d3a2f8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1162.810202] env[62965]: DEBUG oslo_concurrency.lockutils [req-412dbbab-dea3-491d-909d-b5bde3761e09 req-16bb6662-06a2-4a30-a6cf-056c0b84ac0c service nova] Lock "79a4df85-7cb0-4ace-ad29-faec14d3a2f8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1162.810387] env[62965]: DEBUG oslo_concurrency.lockutils [req-412dbbab-dea3-491d-909d-b5bde3761e09 req-16bb6662-06a2-4a30-a6cf-056c0b84ac0c service nova] Lock "79a4df85-7cb0-4ace-ad29-faec14d3a2f8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1162.810556] env[62965]: DEBUG nova.compute.manager [req-412dbbab-dea3-491d-909d-b5bde3761e09 req-16bb6662-06a2-4a30-a6cf-056c0b84ac0c service nova] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] No waiting events found dispatching network-vif-plugged-c265581c-e8f5-4b97-a87c-3c686b21e331 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1162.810720] env[62965]: WARNING nova.compute.manager [req-412dbbab-dea3-491d-909d-b5bde3761e09 req-16bb6662-06a2-4a30-a6cf-056c0b84ac0c service nova] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Received unexpected event network-vif-plugged-c265581c-e8f5-4b97-a87c-3c686b21e331 for instance with vm_state building and task_state spawning. [ 1162.890305] env[62965]: DEBUG nova.network.neutron [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Successfully updated port: c265581c-e8f5-4b97-a87c-3c686b21e331 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1163.207512] env[62965]: DEBUG nova.scheduler.client.report [None req-49059b49-a8ba-445a-9edd-b4268975065b tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1163.713194] env[62965]: DEBUG oslo_concurrency.lockutils [None req-49059b49-a8ba-445a-9edd-b4268975065b tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.620s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1163.730605] env[62965]: INFO nova.scheduler.client.report [None req-49059b49-a8ba-445a-9edd-b4268975065b tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Deleted allocations for instance 385648d7-20e3-442c-b6a1-3d972d266487 [ 1164.239800] env[62965]: DEBUG oslo_concurrency.lockutils [None req-49059b49-a8ba-445a-9edd-b4268975065b tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "385648d7-20e3-442c-b6a1-3d972d266487" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.276s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1164.832197] env[62965]: DEBUG nova.compute.manager [req-9073ca2e-08b7-4723-9cfe-e2d5a26fd46f req-a1dc1af7-ab61-4c42-baf3-711f9cb46e09 service nova] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Received event network-changed-c265581c-e8f5-4b97-a87c-3c686b21e331 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1164.832456] env[62965]: DEBUG nova.compute.manager [req-9073ca2e-08b7-4723-9cfe-e2d5a26fd46f req-a1dc1af7-ab61-4c42-baf3-711f9cb46e09 service nova] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Refreshing instance network info cache due to event network-changed-c265581c-e8f5-4b97-a87c-3c686b21e331. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 1164.832613] env[62965]: DEBUG oslo_concurrency.lockutils [req-9073ca2e-08b7-4723-9cfe-e2d5a26fd46f req-a1dc1af7-ab61-4c42-baf3-711f9cb46e09 service nova] Acquiring lock "refresh_cache-79a4df85-7cb0-4ace-ad29-faec14d3a2f8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1164.832759] env[62965]: DEBUG oslo_concurrency.lockutils [req-9073ca2e-08b7-4723-9cfe-e2d5a26fd46f req-a1dc1af7-ab61-4c42-baf3-711f9cb46e09 service nova] Acquired lock "refresh_cache-79a4df85-7cb0-4ace-ad29-faec14d3a2f8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1164.834724] env[62965]: DEBUG nova.network.neutron [req-9073ca2e-08b7-4723-9cfe-e2d5a26fd46f req-a1dc1af7-ab61-4c42-baf3-711f9cb46e09 service nova] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Refreshing network info cache for port c265581c-e8f5-4b97-a87c-3c686b21e331 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1164.911215] env[62965]: DEBUG nova.compute.manager [req-d960e9f1-7da6-45f2-954c-89698c3084f9 req-273f12b0-5d55-40ee-b7b6-685af2db5282 service nova] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Received event network-vif-plugged-a662d925-10f0-40be-bc51-13a18e973e25 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1164.911428] env[62965]: DEBUG oslo_concurrency.lockutils [req-d960e9f1-7da6-45f2-954c-89698c3084f9 req-273f12b0-5d55-40ee-b7b6-685af2db5282 service nova] Acquiring lock "79a4df85-7cb0-4ace-ad29-faec14d3a2f8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1164.911630] env[62965]: DEBUG oslo_concurrency.lockutils [req-d960e9f1-7da6-45f2-954c-89698c3084f9 req-273f12b0-5d55-40ee-b7b6-685af2db5282 service nova] Lock "79a4df85-7cb0-4ace-ad29-faec14d3a2f8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1164.911845] env[62965]: DEBUG oslo_concurrency.lockutils [req-d960e9f1-7da6-45f2-954c-89698c3084f9 req-273f12b0-5d55-40ee-b7b6-685af2db5282 service nova] Lock "79a4df85-7cb0-4ace-ad29-faec14d3a2f8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1164.911954] env[62965]: DEBUG nova.compute.manager [req-d960e9f1-7da6-45f2-954c-89698c3084f9 req-273f12b0-5d55-40ee-b7b6-685af2db5282 service nova] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] No waiting events found dispatching network-vif-plugged-a662d925-10f0-40be-bc51-13a18e973e25 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1164.912132] env[62965]: WARNING nova.compute.manager [req-d960e9f1-7da6-45f2-954c-89698c3084f9 req-273f12b0-5d55-40ee-b7b6-685af2db5282 service nova] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Received unexpected event network-vif-plugged-a662d925-10f0-40be-bc51-13a18e973e25 for instance with vm_state building and task_state spawning. [ 1165.173712] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "62b5855f-90f6-4480-bbbf-106e1ce6ae2d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1165.173954] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "62b5855f-90f6-4480-bbbf-106e1ce6ae2d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.375683] env[62965]: DEBUG nova.network.neutron [req-9073ca2e-08b7-4723-9cfe-e2d5a26fd46f req-a1dc1af7-ab61-4c42-baf3-711f9cb46e09 service nova] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1165.476048] env[62965]: DEBUG nova.network.neutron [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Successfully updated port: a662d925-10f0-40be-bc51-13a18e973e25 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1165.489780] env[62965]: DEBUG nova.network.neutron [req-9073ca2e-08b7-4723-9cfe-e2d5a26fd46f req-a1dc1af7-ab61-4c42-baf3-711f9cb46e09 service nova] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1165.677089] env[62965]: DEBUG nova.compute.manager [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1165.979141] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Acquiring lock "refresh_cache-79a4df85-7cb0-4ace-ad29-faec14d3a2f8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1165.992816] env[62965]: DEBUG oslo_concurrency.lockutils [req-9073ca2e-08b7-4723-9cfe-e2d5a26fd46f req-a1dc1af7-ab61-4c42-baf3-711f9cb46e09 service nova] Releasing lock "refresh_cache-79a4df85-7cb0-4ace-ad29-faec14d3a2f8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1165.993251] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Acquired lock "refresh_cache-79a4df85-7cb0-4ace-ad29-faec14d3a2f8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1165.993410] env[62965]: DEBUG nova.network.neutron [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1166.200840] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1166.201165] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1166.202648] env[62965]: INFO nova.compute.claims [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1166.527104] env[62965]: DEBUG nova.network.neutron [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1166.775914] env[62965]: DEBUG nova.network.neutron [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Updating instance_info_cache with network_info: [{"id": "c265581c-e8f5-4b97-a87c-3c686b21e331", "address": "fa:16:3e:66:9a:e4", "network": {"id": "6c792c7d-fdf1-4f68-9140-00e6137c82af", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1238158531", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.155", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a757887b3e1e4c69bea03f4924fc813e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc265581c-e8", "ovs_interfaceid": "c265581c-e8f5-4b97-a87c-3c686b21e331", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a662d925-10f0-40be-bc51-13a18e973e25", "address": "fa:16:3e:74:b2:8b", "network": {"id": "4ad14a55-8878-4a8a-91aa-d91feb609eef", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-945632090", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.47", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "a757887b3e1e4c69bea03f4924fc813e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a407774d-9c2a-411d-9d6f-9ca733b97f3f", "external-id": "nsx-vlan-transportzone-710", "segmentation_id": 710, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa662d925-10", "ovs_interfaceid": "a662d925-10f0-40be-bc51-13a18e973e25", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1166.858847] env[62965]: DEBUG nova.compute.manager [req-f149fd3b-d830-4779-813d-c6bebc950cb7 req-08e86ea2-c168-439a-8104-3a2f5b0e983b service nova] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Received event network-changed-a662d925-10f0-40be-bc51-13a18e973e25 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1166.859064] env[62965]: DEBUG nova.compute.manager [req-f149fd3b-d830-4779-813d-c6bebc950cb7 req-08e86ea2-c168-439a-8104-3a2f5b0e983b service nova] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Refreshing instance network info cache due to event network-changed-a662d925-10f0-40be-bc51-13a18e973e25. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 1166.859263] env[62965]: DEBUG oslo_concurrency.lockutils [req-f149fd3b-d830-4779-813d-c6bebc950cb7 req-08e86ea2-c168-439a-8104-3a2f5b0e983b service nova] Acquiring lock "refresh_cache-79a4df85-7cb0-4ace-ad29-faec14d3a2f8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1167.255038] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3d1659f-57f8-4c7e-bc2e-26bd04685c29 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.262299] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d61e75d-7722-444d-8c53-be562d999c5c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.291974] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Releasing lock "refresh_cache-79a4df85-7cb0-4ace-ad29-faec14d3a2f8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1167.292297] env[62965]: DEBUG nova.compute.manager [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Instance network_info: |[{"id": "c265581c-e8f5-4b97-a87c-3c686b21e331", "address": "fa:16:3e:66:9a:e4", "network": {"id": "6c792c7d-fdf1-4f68-9140-00e6137c82af", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1238158531", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.155", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a757887b3e1e4c69bea03f4924fc813e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc265581c-e8", "ovs_interfaceid": "c265581c-e8f5-4b97-a87c-3c686b21e331", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a662d925-10f0-40be-bc51-13a18e973e25", "address": "fa:16:3e:74:b2:8b", "network": {"id": "4ad14a55-8878-4a8a-91aa-d91feb609eef", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-945632090", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.47", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "a757887b3e1e4c69bea03f4924fc813e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a407774d-9c2a-411d-9d6f-9ca733b97f3f", "external-id": "nsx-vlan-transportzone-710", "segmentation_id": 710, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa662d925-10", "ovs_interfaceid": "a662d925-10f0-40be-bc51-13a18e973e25", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1167.292728] env[62965]: DEBUG oslo_concurrency.lockutils [req-f149fd3b-d830-4779-813d-c6bebc950cb7 req-08e86ea2-c168-439a-8104-3a2f5b0e983b service nova] Acquired lock "refresh_cache-79a4df85-7cb0-4ace-ad29-faec14d3a2f8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1167.292995] env[62965]: DEBUG nova.network.neutron [req-f149fd3b-d830-4779-813d-c6bebc950cb7 req-08e86ea2-c168-439a-8104-3a2f5b0e983b service nova] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Refreshing network info cache for port a662d925-10f0-40be-bc51-13a18e973e25 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1167.294255] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:66:9a:e4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e1c9bb98-73a9-48eb-856e-a541afe9b07b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c265581c-e8f5-4b97-a87c-3c686b21e331', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:74:b2:8b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a407774d-9c2a-411d-9d6f-9ca733b97f3f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a662d925-10f0-40be-bc51-13a18e973e25', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1167.303260] env[62965]: DEBUG oslo.service.loopingcall [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1167.303928] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58d75af1-b674-47f2-b15b-34872a31f626 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.308563] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1167.309213] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e1813c4f-9793-401b-8bd8-45764a90c648 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.329644] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b2bf2ad-204f-4475-9f2d-b575652dbb10 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.334222] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1167.334222] env[62965]: value = "task-1390457" [ 1167.334222] env[62965]: _type = "Task" [ 1167.334222] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.345197] env[62965]: DEBUG nova.compute.provider_tree [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1167.350804] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390457, 'name': CreateVM_Task} progress is 6%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.511890] env[62965]: DEBUG nova.network.neutron [req-f149fd3b-d830-4779-813d-c6bebc950cb7 req-08e86ea2-c168-439a-8104-3a2f5b0e983b service nova] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Updated VIF entry in instance network info cache for port a662d925-10f0-40be-bc51-13a18e973e25. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1167.512388] env[62965]: DEBUG nova.network.neutron [req-f149fd3b-d830-4779-813d-c6bebc950cb7 req-08e86ea2-c168-439a-8104-3a2f5b0e983b service nova] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Updating instance_info_cache with network_info: [{"id": "c265581c-e8f5-4b97-a87c-3c686b21e331", "address": "fa:16:3e:66:9a:e4", "network": {"id": "6c792c7d-fdf1-4f68-9140-00e6137c82af", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1238158531", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.155", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a757887b3e1e4c69bea03f4924fc813e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e1c9bb98-73a9-48eb-856e-a541afe9b07b", "external-id": "nsx-vlan-transportzone-755", "segmentation_id": 755, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc265581c-e8", "ovs_interfaceid": "c265581c-e8f5-4b97-a87c-3c686b21e331", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a662d925-10f0-40be-bc51-13a18e973e25", "address": "fa:16:3e:74:b2:8b", "network": {"id": "4ad14a55-8878-4a8a-91aa-d91feb609eef", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-945632090", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.47", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "a757887b3e1e4c69bea03f4924fc813e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a407774d-9c2a-411d-9d6f-9ca733b97f3f", "external-id": "nsx-vlan-transportzone-710", "segmentation_id": 710, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa662d925-10", "ovs_interfaceid": "a662d925-10f0-40be-bc51-13a18e973e25", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1167.844902] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390457, 'name': CreateVM_Task, 'duration_secs': 0.371882} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.845128] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1167.845920] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1167.846168] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1167.846533] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1167.846820] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-18d7ce2b-e030-4218-923a-4ce88df2f71d {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.848867] env[62965]: DEBUG nova.scheduler.client.report [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1167.854330] env[62965]: DEBUG oslo_vmware.api [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Waiting for the task: (returnval){ [ 1167.854330] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52a4d447-12b5-3607-cfac-e97789711e5d" [ 1167.854330] env[62965]: _type = "Task" [ 1167.854330] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.861429] env[62965]: DEBUG oslo_vmware.api [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52a4d447-12b5-3607-cfac-e97789711e5d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.015899] env[62965]: DEBUG oslo_concurrency.lockutils [req-f149fd3b-d830-4779-813d-c6bebc950cb7 req-08e86ea2-c168-439a-8104-3a2f5b0e983b service nova] Releasing lock "refresh_cache-79a4df85-7cb0-4ace-ad29-faec14d3a2f8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1168.353958] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.153s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1168.354538] env[62965]: DEBUG nova.compute.manager [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1168.368109] env[62965]: DEBUG oslo_vmware.api [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52a4d447-12b5-3607-cfac-e97789711e5d, 'name': SearchDatastore_Task, 'duration_secs': 0.008945} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.368379] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1168.368602] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1168.368870] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1168.369042] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1168.369235] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1168.369483] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5618c32f-d8e8-4310-9555-8beac8f75219 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.377279] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1168.377452] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1168.378424] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b0bfd733-b05d-4698-8020-246a8f9f0e2b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.383899] env[62965]: DEBUG oslo_vmware.api [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Waiting for the task: (returnval){ [ 1168.383899] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52fede67-21dc-c9e4-5f7d-4f5573a7ccca" [ 1168.383899] env[62965]: _type = "Task" [ 1168.383899] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.391132] env[62965]: DEBUG oslo_vmware.api [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52fede67-21dc-c9e4-5f7d-4f5573a7ccca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.864502] env[62965]: DEBUG nova.compute.utils [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1168.865974] env[62965]: DEBUG nova.compute.manager [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Allocating IP information in the background. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1168.866179] env[62965]: DEBUG nova.network.neutron [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] allocate_for_instance() {{(pid=62965) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1168.893246] env[62965]: DEBUG oslo_vmware.api [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52fede67-21dc-c9e4-5f7d-4f5573a7ccca, 'name': SearchDatastore_Task, 'duration_secs': 0.008502} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.893980] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ee12c84-2909-4994-b772-b3dce635aa86 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.898525] env[62965]: DEBUG oslo_vmware.api [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Waiting for the task: (returnval){ [ 1168.898525] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52be8653-9bb5-d550-073d-ec4d66d19af1" [ 1168.898525] env[62965]: _type = "Task" [ 1168.898525] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.902399] env[62965]: DEBUG nova.policy [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e9d9f79c936143f5815fcb99755bbd4e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8fab418e2fe1420793517663574b43bb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62965) authorize /opt/stack/nova/nova/policy.py:201}} [ 1168.907724] env[62965]: DEBUG oslo_vmware.api [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52be8653-9bb5-d550-073d-ec4d66d19af1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.149905] env[62965]: DEBUG nova.network.neutron [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Successfully created port: 416c6e70-3506-44e8-ab05-52aef45d2195 {{(pid=62965) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1169.368742] env[62965]: DEBUG nova.compute.manager [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1169.408523] env[62965]: DEBUG oslo_vmware.api [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52be8653-9bb5-d550-073d-ec4d66d19af1, 'name': SearchDatastore_Task, 'duration_secs': 0.009837} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.409457] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1169.409776] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] 79a4df85-7cb0-4ace-ad29-faec14d3a2f8/79a4df85-7cb0-4ace-ad29-faec14d3a2f8.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1169.410086] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cfb9332a-1415-4e7d-a2e4-d7a2d96b6414 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.416989] env[62965]: DEBUG oslo_vmware.api [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Waiting for the task: (returnval){ [ 1169.416989] env[62965]: value = "task-1390458" [ 1169.416989] env[62965]: _type = "Task" [ 1169.416989] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.424864] env[62965]: DEBUG oslo_vmware.api [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': task-1390458, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.821022] env[62965]: DEBUG oslo_concurrency.lockutils [None req-92334b9e-6c97-474b-8bb2-1774cd4bb3b2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquiring lock "73a7ca0e-b1d3-4464-969b-5111d2506a8b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.821361] env[62965]: DEBUG oslo_concurrency.lockutils [None req-92334b9e-6c97-474b-8bb2-1774cd4bb3b2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "73a7ca0e-b1d3-4464-969b-5111d2506a8b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1169.822045] env[62965]: DEBUG oslo_concurrency.lockutils [None req-92334b9e-6c97-474b-8bb2-1774cd4bb3b2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquiring lock "73a7ca0e-b1d3-4464-969b-5111d2506a8b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.822283] env[62965]: DEBUG oslo_concurrency.lockutils [None req-92334b9e-6c97-474b-8bb2-1774cd4bb3b2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "73a7ca0e-b1d3-4464-969b-5111d2506a8b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1169.822478] env[62965]: DEBUG oslo_concurrency.lockutils [None req-92334b9e-6c97-474b-8bb2-1774cd4bb3b2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "73a7ca0e-b1d3-4464-969b-5111d2506a8b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1169.824829] env[62965]: INFO nova.compute.manager [None req-92334b9e-6c97-474b-8bb2-1774cd4bb3b2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Terminating instance [ 1169.928153] env[62965]: DEBUG oslo_vmware.api [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': task-1390458, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.328740] env[62965]: DEBUG nova.compute.manager [None req-92334b9e-6c97-474b-8bb2-1774cd4bb3b2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1170.328969] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-92334b9e-6c97-474b-8bb2-1774cd4bb3b2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1170.329892] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b2e3b90-a5b1-43bb-9766-377983404e26 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.338908] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-92334b9e-6c97-474b-8bb2-1774cd4bb3b2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1170.339154] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d9298849-12af-4073-862d-602ae0c2f0c9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.344667] env[62965]: DEBUG oslo_vmware.api [None req-92334b9e-6c97-474b-8bb2-1774cd4bb3b2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 1170.344667] env[62965]: value = "task-1390459" [ 1170.344667] env[62965]: _type = "Task" [ 1170.344667] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.352016] env[62965]: DEBUG oslo_vmware.api [None req-92334b9e-6c97-474b-8bb2-1774cd4bb3b2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390459, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.379511] env[62965]: DEBUG nova.compute.manager [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1170.405356] env[62965]: DEBUG nova.virt.hardware [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1170.405631] env[62965]: DEBUG nova.virt.hardware [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1170.405791] env[62965]: DEBUG nova.virt.hardware [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1170.405971] env[62965]: DEBUG nova.virt.hardware [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1170.406159] env[62965]: DEBUG nova.virt.hardware [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1170.406314] env[62965]: DEBUG nova.virt.hardware [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1170.406520] env[62965]: DEBUG nova.virt.hardware [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1170.406685] env[62965]: DEBUG nova.virt.hardware [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1170.406910] env[62965]: DEBUG nova.virt.hardware [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1170.407229] env[62965]: DEBUG nova.virt.hardware [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1170.407504] env[62965]: DEBUG nova.virt.hardware [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1170.408525] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6479a5ff-38cc-44b0-b120-32290de25191 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.417136] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e67b489e-127f-4d3b-9448-445f7f026e5c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.436620] env[62965]: DEBUG oslo_vmware.api [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': task-1390458, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.744537} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.436862] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] 79a4df85-7cb0-4ace-ad29-faec14d3a2f8/79a4df85-7cb0-4ace-ad29-faec14d3a2f8.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1170.437138] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1170.437395] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-59f0f3c4-6990-485a-8cc5-9960fd04c8e2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.443754] env[62965]: DEBUG oslo_vmware.api [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Waiting for the task: (returnval){ [ 1170.443754] env[62965]: value = "task-1390460" [ 1170.443754] env[62965]: _type = "Task" [ 1170.443754] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.451791] env[62965]: DEBUG oslo_vmware.api [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': task-1390460, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.522207] env[62965]: DEBUG nova.compute.manager [req-f1837090-58e3-40cb-8210-3c81080f3e9b req-8361e160-557f-4685-b90c-63559e78b60c service nova] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Received event network-vif-plugged-416c6e70-3506-44e8-ab05-52aef45d2195 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1170.522453] env[62965]: DEBUG oslo_concurrency.lockutils [req-f1837090-58e3-40cb-8210-3c81080f3e9b req-8361e160-557f-4685-b90c-63559e78b60c service nova] Acquiring lock "62b5855f-90f6-4480-bbbf-106e1ce6ae2d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1170.522651] env[62965]: DEBUG oslo_concurrency.lockutils [req-f1837090-58e3-40cb-8210-3c81080f3e9b req-8361e160-557f-4685-b90c-63559e78b60c service nova] Lock "62b5855f-90f6-4480-bbbf-106e1ce6ae2d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1170.522793] env[62965]: DEBUG oslo_concurrency.lockutils [req-f1837090-58e3-40cb-8210-3c81080f3e9b req-8361e160-557f-4685-b90c-63559e78b60c service nova] Lock "62b5855f-90f6-4480-bbbf-106e1ce6ae2d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1170.522988] env[62965]: DEBUG nova.compute.manager [req-f1837090-58e3-40cb-8210-3c81080f3e9b req-8361e160-557f-4685-b90c-63559e78b60c service nova] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] No waiting events found dispatching network-vif-plugged-416c6e70-3506-44e8-ab05-52aef45d2195 {{(pid=62965) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1170.523174] env[62965]: WARNING nova.compute.manager [req-f1837090-58e3-40cb-8210-3c81080f3e9b req-8361e160-557f-4685-b90c-63559e78b60c service nova] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Received unexpected event network-vif-plugged-416c6e70-3506-44e8-ab05-52aef45d2195 for instance with vm_state building and task_state spawning. [ 1170.854156] env[62965]: DEBUG oslo_vmware.api [None req-92334b9e-6c97-474b-8bb2-1774cd4bb3b2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390459, 'name': PowerOffVM_Task, 'duration_secs': 0.202272} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.854395] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-92334b9e-6c97-474b-8bb2-1774cd4bb3b2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1170.854562] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-92334b9e-6c97-474b-8bb2-1774cd4bb3b2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1170.854809] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-17d491b7-ce30-4865-bfb0-f72fd15c0cf4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.915207] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-92334b9e-6c97-474b-8bb2-1774cd4bb3b2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1170.915474] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-92334b9e-6c97-474b-8bb2-1774cd4bb3b2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Deleting contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1170.915671] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-92334b9e-6c97-474b-8bb2-1774cd4bb3b2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Deleting the datastore file [datastore2] 73a7ca0e-b1d3-4464-969b-5111d2506a8b {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1170.915944] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9980a4f2-f8b3-4727-b6d1-ab6e017ce898 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.921410] env[62965]: DEBUG oslo_vmware.api [None req-92334b9e-6c97-474b-8bb2-1774cd4bb3b2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for the task: (returnval){ [ 1170.921410] env[62965]: value = "task-1390462" [ 1170.921410] env[62965]: _type = "Task" [ 1170.921410] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.929056] env[62965]: DEBUG oslo_vmware.api [None req-92334b9e-6c97-474b-8bb2-1774cd4bb3b2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390462, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.954637] env[62965]: DEBUG oslo_vmware.api [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': task-1390460, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064215} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.954899] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1170.955710] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18fe1fb0-a632-4a9d-b93a-2bdec9c9ff8a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.979554] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] 79a4df85-7cb0-4ace-ad29-faec14d3a2f8/79a4df85-7cb0-4ace-ad29-faec14d3a2f8.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1170.979829] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c5c9988f-5dd6-4a7b-b268-7be5a81b696b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.000103] env[62965]: DEBUG oslo_vmware.api [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Waiting for the task: (returnval){ [ 1171.000103] env[62965]: value = "task-1390463" [ 1171.000103] env[62965]: _type = "Task" [ 1171.000103] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.008048] env[62965]: DEBUG oslo_vmware.api [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': task-1390463, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.088213] env[62965]: DEBUG nova.network.neutron [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Successfully updated port: 416c6e70-3506-44e8-ab05-52aef45d2195 {{(pid=62965) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1171.136191] env[62965]: DEBUG nova.compute.manager [req-e9740afa-83f3-4309-b06a-6a95b7d1baef req-1d68e2ef-661f-4f4a-aa1d-f10569ed2e57 service nova] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Received event network-changed-416c6e70-3506-44e8-ab05-52aef45d2195 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1171.136191] env[62965]: DEBUG nova.compute.manager [req-e9740afa-83f3-4309-b06a-6a95b7d1baef req-1d68e2ef-661f-4f4a-aa1d-f10569ed2e57 service nova] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Refreshing instance network info cache due to event network-changed-416c6e70-3506-44e8-ab05-52aef45d2195. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 1171.136394] env[62965]: DEBUG oslo_concurrency.lockutils [req-e9740afa-83f3-4309-b06a-6a95b7d1baef req-1d68e2ef-661f-4f4a-aa1d-f10569ed2e57 service nova] Acquiring lock "refresh_cache-62b5855f-90f6-4480-bbbf-106e1ce6ae2d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1171.136524] env[62965]: DEBUG oslo_concurrency.lockutils [req-e9740afa-83f3-4309-b06a-6a95b7d1baef req-1d68e2ef-661f-4f4a-aa1d-f10569ed2e57 service nova] Acquired lock "refresh_cache-62b5855f-90f6-4480-bbbf-106e1ce6ae2d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1171.136689] env[62965]: DEBUG nova.network.neutron [req-e9740afa-83f3-4309-b06a-6a95b7d1baef req-1d68e2ef-661f-4f4a-aa1d-f10569ed2e57 service nova] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Refreshing network info cache for port 416c6e70-3506-44e8-ab05-52aef45d2195 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1171.432488] env[62965]: DEBUG oslo_vmware.api [None req-92334b9e-6c97-474b-8bb2-1774cd4bb3b2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Task: {'id': task-1390462, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.192919} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.432488] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-92334b9e-6c97-474b-8bb2-1774cd4bb3b2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1171.432861] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-92334b9e-6c97-474b-8bb2-1774cd4bb3b2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Deleted contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1171.432861] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-92334b9e-6c97-474b-8bb2-1774cd4bb3b2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1171.433140] env[62965]: INFO nova.compute.manager [None req-92334b9e-6c97-474b-8bb2-1774cd4bb3b2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1171.433447] env[62965]: DEBUG oslo.service.loopingcall [None req-92334b9e-6c97-474b-8bb2-1774cd4bb3b2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1171.433692] env[62965]: DEBUG nova.compute.manager [-] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1171.433821] env[62965]: DEBUG nova.network.neutron [-] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1171.510463] env[62965]: DEBUG oslo_vmware.api [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': task-1390463, 'name': ReconfigVM_Task, 'duration_secs': 0.498478} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.510744] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Reconfigured VM instance instance-0000006d to attach disk [datastore1] 79a4df85-7cb0-4ace-ad29-faec14d3a2f8/79a4df85-7cb0-4ace-ad29-faec14d3a2f8.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1171.511392] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a351a713-bdef-410f-ba34-c27267121239 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.517594] env[62965]: DEBUG oslo_vmware.api [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Waiting for the task: (returnval){ [ 1171.517594] env[62965]: value = "task-1390464" [ 1171.517594] env[62965]: _type = "Task" [ 1171.517594] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.525203] env[62965]: DEBUG oslo_vmware.api [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': task-1390464, 'name': Rename_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.591183] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "refresh_cache-62b5855f-90f6-4480-bbbf-106e1ce6ae2d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1171.672620] env[62965]: DEBUG nova.network.neutron [req-e9740afa-83f3-4309-b06a-6a95b7d1baef req-1d68e2ef-661f-4f4a-aa1d-f10569ed2e57 service nova] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1171.763103] env[62965]: DEBUG nova.network.neutron [req-e9740afa-83f3-4309-b06a-6a95b7d1baef req-1d68e2ef-661f-4f4a-aa1d-f10569ed2e57 service nova] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1172.027622] env[62965]: DEBUG oslo_vmware.api [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': task-1390464, 'name': Rename_Task, 'duration_secs': 0.128615} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.027898] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1172.028233] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-96395a52-8e52-412f-9544-ec8953d0adb6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.034576] env[62965]: DEBUG oslo_vmware.api [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Waiting for the task: (returnval){ [ 1172.034576] env[62965]: value = "task-1390465" [ 1172.034576] env[62965]: _type = "Task" [ 1172.034576] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.042922] env[62965]: DEBUG oslo_vmware.api [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': task-1390465, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.265648] env[62965]: DEBUG oslo_concurrency.lockutils [req-e9740afa-83f3-4309-b06a-6a95b7d1baef req-1d68e2ef-661f-4f4a-aa1d-f10569ed2e57 service nova] Releasing lock "refresh_cache-62b5855f-90f6-4480-bbbf-106e1ce6ae2d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1172.266042] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquired lock "refresh_cache-62b5855f-90f6-4480-bbbf-106e1ce6ae2d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1172.266204] env[62965]: DEBUG nova.network.neutron [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1172.363089] env[62965]: DEBUG nova.network.neutron [-] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1172.546046] env[62965]: DEBUG oslo_vmware.api [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': task-1390465, 'name': PowerOnVM_Task, 'duration_secs': 0.432263} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.546046] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1172.546046] env[62965]: INFO nova.compute.manager [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Took 10.10 seconds to spawn the instance on the hypervisor. [ 1172.546046] env[62965]: DEBUG nova.compute.manager [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1172.546856] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9d155d7-c9c4-4576-8d23-4e96d3713d58 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.803951] env[62965]: DEBUG nova.network.neutron [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1172.865953] env[62965]: INFO nova.compute.manager [-] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Took 1.43 seconds to deallocate network for instance. [ 1172.939327] env[62965]: DEBUG nova.network.neutron [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Updating instance_info_cache with network_info: [{"id": "416c6e70-3506-44e8-ab05-52aef45d2195", "address": "fa:16:3e:cb:c6:e5", "network": {"id": "4649fed7-c2ca-43a9-a183-7c99423da726", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1978698173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fab418e2fe1420793517663574b43bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43ad01d2-c7dd-453c-a929-8ad76294d13c", "external-id": "nsx-vlan-transportzone-176", "segmentation_id": 176, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap416c6e70-35", "ovs_interfaceid": "416c6e70-3506-44e8-ab05-52aef45d2195", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1173.063206] env[62965]: INFO nova.compute.manager [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Took 16.02 seconds to build instance. [ 1173.159937] env[62965]: DEBUG nova.compute.manager [req-e5f815e2-fc79-44fe-ac88-48bc6f7e0335 req-648dc5dc-7891-4786-b5c6-f90a52b30e37 service nova] [instance: 73a7ca0e-b1d3-4464-969b-5111d2506a8b] Received event network-vif-deleted-5f939a36-bc5f-4fd7-b29e-011f72b3256c {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1173.372397] env[62965]: DEBUG oslo_concurrency.lockutils [None req-92334b9e-6c97-474b-8bb2-1774cd4bb3b2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1173.372670] env[62965]: DEBUG oslo_concurrency.lockutils [None req-92334b9e-6c97-474b-8bb2-1774cd4bb3b2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1173.372889] env[62965]: DEBUG nova.objects.instance [None req-92334b9e-6c97-474b-8bb2-1774cd4bb3b2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lazy-loading 'resources' on Instance uuid 73a7ca0e-b1d3-4464-969b-5111d2506a8b {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1173.442594] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Releasing lock "refresh_cache-62b5855f-90f6-4480-bbbf-106e1ce6ae2d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1173.442900] env[62965]: DEBUG nova.compute.manager [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Instance network_info: |[{"id": "416c6e70-3506-44e8-ab05-52aef45d2195", "address": "fa:16:3e:cb:c6:e5", "network": {"id": "4649fed7-c2ca-43a9-a183-7c99423da726", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1978698173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fab418e2fe1420793517663574b43bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43ad01d2-c7dd-453c-a929-8ad76294d13c", "external-id": "nsx-vlan-transportzone-176", "segmentation_id": 176, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap416c6e70-35", "ovs_interfaceid": "416c6e70-3506-44e8-ab05-52aef45d2195", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1173.443396] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cb:c6:e5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '43ad01d2-c7dd-453c-a929-8ad76294d13c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '416c6e70-3506-44e8-ab05-52aef45d2195', 'vif_model': 'vmxnet3'}] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1173.450799] env[62965]: DEBUG oslo.service.loopingcall [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1173.451258] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1173.451497] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d1c9d5a4-6f6e-42f7-8587-acbb86be13ce {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.471858] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1173.471858] env[62965]: value = "task-1390466" [ 1173.471858] env[62965]: _type = "Task" [ 1173.471858] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.479111] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390466, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.565146] env[62965]: DEBUG oslo_concurrency.lockutils [None req-b4e81472-e042-4979-bae8-e9c778aee56a tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Lock "79a4df85-7cb0-4ace-ad29-faec14d3a2f8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.535s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1173.791599] env[62965]: DEBUG oslo_concurrency.lockutils [None req-dc607bc3-40ec-4198-a4c5-1b0f7235bb17 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Acquiring lock "79a4df85-7cb0-4ace-ad29-faec14d3a2f8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1173.791887] env[62965]: DEBUG oslo_concurrency.lockutils [None req-dc607bc3-40ec-4198-a4c5-1b0f7235bb17 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Lock "79a4df85-7cb0-4ace-ad29-faec14d3a2f8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1173.792133] env[62965]: DEBUG oslo_concurrency.lockutils [None req-dc607bc3-40ec-4198-a4c5-1b0f7235bb17 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Acquiring lock "79a4df85-7cb0-4ace-ad29-faec14d3a2f8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1173.792328] env[62965]: DEBUG oslo_concurrency.lockutils [None req-dc607bc3-40ec-4198-a4c5-1b0f7235bb17 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Lock "79a4df85-7cb0-4ace-ad29-faec14d3a2f8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1173.792504] env[62965]: DEBUG oslo_concurrency.lockutils [None req-dc607bc3-40ec-4198-a4c5-1b0f7235bb17 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Lock "79a4df85-7cb0-4ace-ad29-faec14d3a2f8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1173.794710] env[62965]: INFO nova.compute.manager [None req-dc607bc3-40ec-4198-a4c5-1b0f7235bb17 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Terminating instance [ 1173.925060] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ab76599-9c8d-47d2-9c77-a747ab5b50fa {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.932484] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be8ee3eb-4e08-46b6-8466-78fc20fdb210 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.962227] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b708fe84-b1f9-410f-bd35-8bba08ccfca6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.969125] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83b7be88-0101-43ea-bce9-e8cd95c0afb8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.986271] env[62965]: DEBUG nova.compute.provider_tree [None req-92334b9e-6c97-474b-8bb2-1774cd4bb3b2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1173.989865] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390466, 'name': CreateVM_Task, 'duration_secs': 0.29103} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.990157] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1173.990765] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1173.990924] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1173.991263] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1173.991484] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce2c9c81-69dd-493d-b72a-626f3a693c29 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.995920] env[62965]: DEBUG oslo_vmware.api [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1173.995920] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52916f55-3690-a11d-1f31-a56de098d0ed" [ 1173.995920] env[62965]: _type = "Task" [ 1173.995920] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.003536] env[62965]: DEBUG oslo_vmware.api [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52916f55-3690-a11d-1f31-a56de098d0ed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.297796] env[62965]: DEBUG nova.compute.manager [None req-dc607bc3-40ec-4198-a4c5-1b0f7235bb17 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1174.298068] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-dc607bc3-40ec-4198-a4c5-1b0f7235bb17 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1174.298951] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f670027-cfc5-4330-94be-05d1f613b934 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.306639] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc607bc3-40ec-4198-a4c5-1b0f7235bb17 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1174.306892] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7dc72d16-697b-4008-af88-f060085959c9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.313116] env[62965]: DEBUG oslo_vmware.api [None req-dc607bc3-40ec-4198-a4c5-1b0f7235bb17 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Waiting for the task: (returnval){ [ 1174.313116] env[62965]: value = "task-1390467" [ 1174.313116] env[62965]: _type = "Task" [ 1174.313116] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.320954] env[62965]: DEBUG oslo_vmware.api [None req-dc607bc3-40ec-4198-a4c5-1b0f7235bb17 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': task-1390467, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.491307] env[62965]: DEBUG nova.scheduler.client.report [None req-92334b9e-6c97-474b-8bb2-1774cd4bb3b2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1174.509431] env[62965]: DEBUG oslo_vmware.api [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52916f55-3690-a11d-1f31-a56de098d0ed, 'name': SearchDatastore_Task, 'duration_secs': 0.009673} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.510624] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1174.510945] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1174.511314] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1174.511531] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquired lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1174.511830] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1174.512428] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5890a4de-daa3-4f7a-b0a6-f633f8a0766f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.521714] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1174.521963] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1174.522951] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66a7e4c6-8c67-4f4f-bb19-73ade549f1d4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.529641] env[62965]: DEBUG oslo_vmware.api [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1174.529641] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52a35b9f-0f92-1d4e-93a5-2a640bf792ad" [ 1174.529641] env[62965]: _type = "Task" [ 1174.529641] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.540750] env[62965]: DEBUG oslo_vmware.api [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52a35b9f-0f92-1d4e-93a5-2a640bf792ad, 'name': SearchDatastore_Task, 'duration_secs': 0.007819} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.541761] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1307ad2c-9156-4c8c-826b-fb560886d402 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.547679] env[62965]: DEBUG oslo_vmware.api [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1174.547679] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]526e6547-c264-cf7d-6959-779ff598dcfd" [ 1174.547679] env[62965]: _type = "Task" [ 1174.547679] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.555407] env[62965]: DEBUG oslo_vmware.api [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]526e6547-c264-cf7d-6959-779ff598dcfd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.823574] env[62965]: DEBUG oslo_vmware.api [None req-dc607bc3-40ec-4198-a4c5-1b0f7235bb17 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': task-1390467, 'name': PowerOffVM_Task, 'duration_secs': 0.193943} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.823944] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc607bc3-40ec-4198-a4c5-1b0f7235bb17 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1174.824074] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-dc607bc3-40ec-4198-a4c5-1b0f7235bb17 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1174.824329] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6e1536d2-8915-4a59-8a72-3238ca19be30 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.926064] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-dc607bc3-40ec-4198-a4c5-1b0f7235bb17 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1174.926295] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-dc607bc3-40ec-4198-a4c5-1b0f7235bb17 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Deleting contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1174.926478] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc607bc3-40ec-4198-a4c5-1b0f7235bb17 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Deleting the datastore file [datastore1] 79a4df85-7cb0-4ace-ad29-faec14d3a2f8 {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1174.926754] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-194f9f97-9738-449c-b996-3a4cd5c59783 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.932533] env[62965]: DEBUG oslo_vmware.api [None req-dc607bc3-40ec-4198-a4c5-1b0f7235bb17 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Waiting for the task: (returnval){ [ 1174.932533] env[62965]: value = "task-1390469" [ 1174.932533] env[62965]: _type = "Task" [ 1174.932533] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.940048] env[62965]: DEBUG oslo_vmware.api [None req-dc607bc3-40ec-4198-a4c5-1b0f7235bb17 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': task-1390469, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.998166] env[62965]: DEBUG oslo_concurrency.lockutils [None req-92334b9e-6c97-474b-8bb2-1774cd4bb3b2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.625s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1175.016967] env[62965]: INFO nova.scheduler.client.report [None req-92334b9e-6c97-474b-8bb2-1774cd4bb3b2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Deleted allocations for instance 73a7ca0e-b1d3-4464-969b-5111d2506a8b [ 1175.058436] env[62965]: DEBUG oslo_vmware.api [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]526e6547-c264-cf7d-6959-779ff598dcfd, 'name': SearchDatastore_Task, 'duration_secs': 0.00736} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.058723] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Releasing lock "[datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1175.059047] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] 62b5855f-90f6-4480-bbbf-106e1ce6ae2d/62b5855f-90f6-4480-bbbf-106e1ce6ae2d.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1175.059895] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-928fe043-738e-427c-b7ae-d4d3c893fb5b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.066267] env[62965]: DEBUG oslo_vmware.api [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1175.066267] env[62965]: value = "task-1390470" [ 1175.066267] env[62965]: _type = "Task" [ 1175.066267] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.075773] env[62965]: DEBUG oslo_vmware.api [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390470, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.444087] env[62965]: DEBUG oslo_vmware.api [None req-dc607bc3-40ec-4198-a4c5-1b0f7235bb17 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Task: {'id': task-1390469, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146095} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.444400] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc607bc3-40ec-4198-a4c5-1b0f7235bb17 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1175.444653] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-dc607bc3-40ec-4198-a4c5-1b0f7235bb17 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Deleted contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1175.444855] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-dc607bc3-40ec-4198-a4c5-1b0f7235bb17 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1175.445153] env[62965]: INFO nova.compute.manager [None req-dc607bc3-40ec-4198-a4c5-1b0f7235bb17 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1175.445539] env[62965]: DEBUG oslo.service.loopingcall [None req-dc607bc3-40ec-4198-a4c5-1b0f7235bb17 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1175.445644] env[62965]: DEBUG nova.compute.manager [-] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1175.445737] env[62965]: DEBUG nova.network.neutron [-] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1175.524916] env[62965]: DEBUG oslo_concurrency.lockutils [None req-92334b9e-6c97-474b-8bb2-1774cd4bb3b2 tempest-AttachVolumeShelveTestJSON-1590987501 tempest-AttachVolumeShelveTestJSON-1590987501-project-member] Lock "73a7ca0e-b1d3-4464-969b-5111d2506a8b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.703s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1175.576334] env[62965]: DEBUG oslo_vmware.api [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390470, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.438544} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.576604] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore1] 62b5855f-90f6-4480-bbbf-106e1ce6ae2d/62b5855f-90f6-4480-bbbf-106e1ce6ae2d.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1175.576818] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1175.577080] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f530801e-8785-42f3-95eb-30813dc1ea2e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.583532] env[62965]: DEBUG oslo_vmware.api [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1175.583532] env[62965]: value = "task-1390471" [ 1175.583532] env[62965]: _type = "Task" [ 1175.583532] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.591121] env[62965]: DEBUG oslo_vmware.api [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390471, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.743138] env[62965]: DEBUG nova.compute.manager [req-07a3b99b-f030-48e5-bb2e-c57e6616f39f req-7cbb1dcc-244c-44d9-8722-dd3bff24d94f service nova] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Received event network-vif-deleted-c265581c-e8f5-4b97-a87c-3c686b21e331 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1175.743375] env[62965]: INFO nova.compute.manager [req-07a3b99b-f030-48e5-bb2e-c57e6616f39f req-7cbb1dcc-244c-44d9-8722-dd3bff24d94f service nova] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Neutron deleted interface c265581c-e8f5-4b97-a87c-3c686b21e331; detaching it from the instance and deleting it from the info cache [ 1175.743624] env[62965]: DEBUG nova.network.neutron [req-07a3b99b-f030-48e5-bb2e-c57e6616f39f req-7cbb1dcc-244c-44d9-8722-dd3bff24d94f service nova] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Updating instance_info_cache with network_info: [{"id": "a662d925-10f0-40be-bc51-13a18e973e25", "address": "fa:16:3e:74:b2:8b", "network": {"id": "4ad14a55-8878-4a8a-91aa-d91feb609eef", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-945632090", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.47", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "a757887b3e1e4c69bea03f4924fc813e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a407774d-9c2a-411d-9d6f-9ca733b97f3f", "external-id": "nsx-vlan-transportzone-710", "segmentation_id": 710, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa662d925-10", "ovs_interfaceid": "a662d925-10f0-40be-bc51-13a18e973e25", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1176.093060] env[62965]: DEBUG oslo_vmware.api [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390471, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.055823} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.093424] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1176.094135] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bce94768-159c-4604-b973-6365ec7318cc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.116053] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] 62b5855f-90f6-4480-bbbf-106e1ce6ae2d/62b5855f-90f6-4480-bbbf-106e1ce6ae2d.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1176.116053] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-51e740c7-76de-469c-aee4-a33ae4151c08 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.135377] env[62965]: DEBUG oslo_vmware.api [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1176.135377] env[62965]: value = "task-1390472" [ 1176.135377] env[62965]: _type = "Task" [ 1176.135377] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.143756] env[62965]: DEBUG oslo_vmware.api [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390472, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.246685] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fd1402d6-74ff-4c58-859f-c8af0ccb6269 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.255019] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-726f8163-5d72-4cd1-9ac7-aa4e95ea8081 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.279537] env[62965]: DEBUG nova.compute.manager [req-07a3b99b-f030-48e5-bb2e-c57e6616f39f req-7cbb1dcc-244c-44d9-8722-dd3bff24d94f service nova] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Detach interface failed, port_id=c265581c-e8f5-4b97-a87c-3c686b21e331, reason: Instance 79a4df85-7cb0-4ace-ad29-faec14d3a2f8 could not be found. {{(pid=62965) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11294}} [ 1176.374385] env[62965]: DEBUG nova.network.neutron [-] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1176.646905] env[62965]: DEBUG oslo_vmware.api [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390472, 'name': ReconfigVM_Task, 'duration_secs': 0.247221} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.646905] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Reconfigured VM instance instance-0000006e to attach disk [datastore1] 62b5855f-90f6-4480-bbbf-106e1ce6ae2d/62b5855f-90f6-4480-bbbf-106e1ce6ae2d.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1176.647110] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b25391a9-765d-46cb-bca8-c6a76383d9f5 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.652716] env[62965]: DEBUG oslo_vmware.api [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1176.652716] env[62965]: value = "task-1390473" [ 1176.652716] env[62965]: _type = "Task" [ 1176.652716] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.665644] env[62965]: DEBUG oslo_vmware.api [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390473, 'name': Rename_Task} progress is 5%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.877291] env[62965]: INFO nova.compute.manager [-] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Took 1.43 seconds to deallocate network for instance. [ 1177.163285] env[62965]: DEBUG oslo_vmware.api [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390473, 'name': Rename_Task, 'duration_secs': 0.132921} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.163646] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1177.163747] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9c76d33b-af58-4599-9221-8e5324aa96de {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.169752] env[62965]: DEBUG oslo_vmware.api [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1177.169752] env[62965]: value = "task-1390474" [ 1177.169752] env[62965]: _type = "Task" [ 1177.169752] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.177614] env[62965]: DEBUG oslo_vmware.api [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390474, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.383400] env[62965]: DEBUG oslo_concurrency.lockutils [None req-dc607bc3-40ec-4198-a4c5-1b0f7235bb17 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1177.383724] env[62965]: DEBUG oslo_concurrency.lockutils [None req-dc607bc3-40ec-4198-a4c5-1b0f7235bb17 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1177.384028] env[62965]: DEBUG nova.objects.instance [None req-dc607bc3-40ec-4198-a4c5-1b0f7235bb17 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Lazy-loading 'resources' on Instance uuid 79a4df85-7cb0-4ace-ad29-faec14d3a2f8 {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1177.679978] env[62965]: DEBUG oslo_vmware.api [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390474, 'name': PowerOnVM_Task, 'duration_secs': 0.40579} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.680373] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1177.680635] env[62965]: INFO nova.compute.manager [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Took 7.30 seconds to spawn the instance on the hypervisor. [ 1177.680862] env[62965]: DEBUG nova.compute.manager [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1177.681746] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3992a885-9b82-48b4-94e4-41d61ffc834e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.766524] env[62965]: DEBUG nova.compute.manager [req-a6bf60bd-caa6-4652-b91c-9981907d1eca req-09c8132a-7961-4c9d-ac04-e2157371a9c9 service nova] [instance: 79a4df85-7cb0-4ace-ad29-faec14d3a2f8] Received event network-vif-deleted-a662d925-10f0-40be-bc51-13a18e973e25 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1177.928935] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6a6ec90-aaaf-4954-af0b-6bbec0a77893 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.936307] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d31e44de-37b7-49dd-912a-a43d39fc1ede {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.966197] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44535083-e504-430b-9c52-c9d69d0f187b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.972966] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90f0e3d5-44f8-4f09-9784-7c21d7330bdb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.986077] env[62965]: DEBUG nova.compute.provider_tree [None req-dc607bc3-40ec-4198-a4c5-1b0f7235bb17 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1178.199103] env[62965]: INFO nova.compute.manager [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Took 12.01 seconds to build instance. [ 1178.489021] env[62965]: DEBUG nova.scheduler.client.report [None req-dc607bc3-40ec-4198-a4c5-1b0f7235bb17 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1178.701209] env[62965]: DEBUG oslo_concurrency.lockutils [None req-2cb74d3f-3413-40c4-9b18-1eb84309d7b0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "62b5855f-90f6-4480-bbbf-106e1ce6ae2d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.527s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1178.994759] env[62965]: DEBUG oslo_concurrency.lockutils [None req-dc607bc3-40ec-4198-a4c5-1b0f7235bb17 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.611s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1179.013387] env[62965]: INFO nova.scheduler.client.report [None req-dc607bc3-40ec-4198-a4c5-1b0f7235bb17 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Deleted allocations for instance 79a4df85-7cb0-4ace-ad29-faec14d3a2f8 [ 1179.285561] env[62965]: DEBUG nova.compute.manager [req-fd4e349e-92d0-469d-abe0-ddf850b28a0d req-8e8ce22d-8d7c-4bd0-99b0-1f1d985756eb service nova] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Received event network-changed-416c6e70-3506-44e8-ab05-52aef45d2195 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1179.285811] env[62965]: DEBUG nova.compute.manager [req-fd4e349e-92d0-469d-abe0-ddf850b28a0d req-8e8ce22d-8d7c-4bd0-99b0-1f1d985756eb service nova] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Refreshing instance network info cache due to event network-changed-416c6e70-3506-44e8-ab05-52aef45d2195. {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11465}} [ 1179.285978] env[62965]: DEBUG oslo_concurrency.lockutils [req-fd4e349e-92d0-469d-abe0-ddf850b28a0d req-8e8ce22d-8d7c-4bd0-99b0-1f1d985756eb service nova] Acquiring lock "refresh_cache-62b5855f-90f6-4480-bbbf-106e1ce6ae2d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1179.286140] env[62965]: DEBUG oslo_concurrency.lockutils [req-fd4e349e-92d0-469d-abe0-ddf850b28a0d req-8e8ce22d-8d7c-4bd0-99b0-1f1d985756eb service nova] Acquired lock "refresh_cache-62b5855f-90f6-4480-bbbf-106e1ce6ae2d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1179.286313] env[62965]: DEBUG nova.network.neutron [req-fd4e349e-92d0-469d-abe0-ddf850b28a0d req-8e8ce22d-8d7c-4bd0-99b0-1f1d985756eb service nova] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Refreshing network info cache for port 416c6e70-3506-44e8-ab05-52aef45d2195 {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1179.521518] env[62965]: DEBUG oslo_concurrency.lockutils [None req-dc607bc3-40ec-4198-a4c5-1b0f7235bb17 tempest-ServersTestMultiNic-2101629131 tempest-ServersTestMultiNic-2101629131-project-member] Lock "79a4df85-7cb0-4ace-ad29-faec14d3a2f8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.729s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1180.177679] env[62965]: DEBUG nova.network.neutron [req-fd4e349e-92d0-469d-abe0-ddf850b28a0d req-8e8ce22d-8d7c-4bd0-99b0-1f1d985756eb service nova] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Updated VIF entry in instance network info cache for port 416c6e70-3506-44e8-ab05-52aef45d2195. {{(pid=62965) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1180.178077] env[62965]: DEBUG nova.network.neutron [req-fd4e349e-92d0-469d-abe0-ddf850b28a0d req-8e8ce22d-8d7c-4bd0-99b0-1f1d985756eb service nova] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Updating instance_info_cache with network_info: [{"id": "416c6e70-3506-44e8-ab05-52aef45d2195", "address": "fa:16:3e:cb:c6:e5", "network": {"id": "4649fed7-c2ca-43a9-a183-7c99423da726", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1978698173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fab418e2fe1420793517663574b43bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43ad01d2-c7dd-453c-a929-8ad76294d13c", "external-id": "nsx-vlan-transportzone-176", "segmentation_id": 176, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap416c6e70-35", "ovs_interfaceid": "416c6e70-3506-44e8-ab05-52aef45d2195", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1180.680765] env[62965]: DEBUG oslo_concurrency.lockutils [req-fd4e349e-92d0-469d-abe0-ddf850b28a0d req-8e8ce22d-8d7c-4bd0-99b0-1f1d985756eb service nova] Releasing lock "refresh_cache-62b5855f-90f6-4480-bbbf-106e1ce6ae2d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1209.104760] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1209.105240] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Starting heal instance info cache {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10257}} [ 1209.608306] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Didn't find any instances for network info cache update. {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10343}} [ 1211.104974] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1212.104629] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1212.104875] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1212.105033] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62965) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10876}} [ 1214.105353] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1215.099189] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1215.103832] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1216.063686] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8292954e-b86e-4f8b-92f6-11eb40a5e6c0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "62b5855f-90f6-4480-bbbf-106e1ce6ae2d" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1216.064059] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8292954e-b86e-4f8b-92f6-11eb40a5e6c0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "62b5855f-90f6-4480-bbbf-106e1ce6ae2d" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1216.064201] env[62965]: DEBUG nova.compute.manager [None req-8292954e-b86e-4f8b-92f6-11eb40a5e6c0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1216.065247] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-292bf4d1-b06e-43e3-974f-d7fa1eef889a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.072320] env[62965]: DEBUG nova.compute.manager [None req-8292954e-b86e-4f8b-92f6-11eb40a5e6c0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62965) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 1216.072879] env[62965]: DEBUG nova.objects.instance [None req-8292954e-b86e-4f8b-92f6-11eb40a5e6c0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lazy-loading 'flavor' on Instance uuid 62b5855f-90f6-4480-bbbf-106e1ce6ae2d {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1216.098596] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1217.080380] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-8292954e-b86e-4f8b-92f6-11eb40a5e6c0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1217.080768] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bf128d1d-16e5-4949-9de1-31e735bdf2f7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.088054] env[62965]: DEBUG oslo_vmware.api [None req-8292954e-b86e-4f8b-92f6-11eb40a5e6c0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1217.088054] env[62965]: value = "task-1390476" [ 1217.088054] env[62965]: _type = "Task" [ 1217.088054] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1217.096652] env[62965]: DEBUG oslo_vmware.api [None req-8292954e-b86e-4f8b-92f6-11eb40a5e6c0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390476, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.104558] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1217.104784] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1217.598207] env[62965]: DEBUG oslo_vmware.api [None req-8292954e-b86e-4f8b-92f6-11eb40a5e6c0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390476, 'name': PowerOffVM_Task, 'duration_secs': 0.169734} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1217.598477] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-8292954e-b86e-4f8b-92f6-11eb40a5e6c0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1217.598683] env[62965]: DEBUG nova.compute.manager [None req-8292954e-b86e-4f8b-92f6-11eb40a5e6c0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1217.599466] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24285758-c303-4559-9c23-cabe0f3c65e9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.607157] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1217.607359] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1217.607521] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1217.607689] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62965) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1217.608586] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb6742c4-0740-48b7-8ab4-72f3e5ce68ec {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.615238] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34607408-9f07-4c0e-a515-80de8e70391a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.629819] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-280edaca-0e84-45f5-af3d-672491f933dc {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.636398] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87ca79ca-ae95-4a7f-af61-1a9e8251ed8b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.665595] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181143MB free_disk=175GB free_vcpus=48 pci_devices=None {{(pid=62965) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1217.665723] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1217.665908] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1218.110735] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8292954e-b86e-4f8b-92f6-11eb40a5e6c0 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "62b5855f-90f6-4480-bbbf-106e1ce6ae2d" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.047s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1218.417092] env[62965]: DEBUG nova.objects.instance [None req-d57256ab-0308-4099-b0de-8a4778d9a1fd tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lazy-loading 'flavor' on Instance uuid 62b5855f-90f6-4480-bbbf-106e1ce6ae2d {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1218.691195] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Instance 62b5855f-90f6-4480-bbbf-106e1ce6ae2d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62965) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1218.691393] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=62965) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1218.691563] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=62965) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1218.716229] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15e30153-48cc-47c0-aed2-98008d5c9853 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.723737] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57568317-810f-48f5-87b6-77e6cd67dff0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.753706] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce3cb3b0-28fc-4ab4-b231-3775ea148057 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.760596] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e631a20-1879-4422-bf4c-30e6e9d94288 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.773351] env[62965]: DEBUG nova.compute.provider_tree [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1218.922060] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d57256ab-0308-4099-b0de-8a4778d9a1fd tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "refresh_cache-62b5855f-90f6-4480-bbbf-106e1ce6ae2d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1218.922280] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d57256ab-0308-4099-b0de-8a4778d9a1fd tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquired lock "refresh_cache-62b5855f-90f6-4480-bbbf-106e1ce6ae2d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1218.922492] env[62965]: DEBUG nova.network.neutron [None req-d57256ab-0308-4099-b0de-8a4778d9a1fd tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1218.922704] env[62965]: DEBUG nova.objects.instance [None req-d57256ab-0308-4099-b0de-8a4778d9a1fd tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lazy-loading 'info_cache' on Instance uuid 62b5855f-90f6-4480-bbbf-106e1ce6ae2d {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1219.276409] env[62965]: DEBUG nova.scheduler.client.report [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1219.426089] env[62965]: DEBUG nova.objects.base [None req-d57256ab-0308-4099-b0de-8a4778d9a1fd tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Object Instance<62b5855f-90f6-4480-bbbf-106e1ce6ae2d> lazy-loaded attributes: flavor,info_cache {{(pid=62965) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1219.781651] env[62965]: DEBUG nova.compute.resource_tracker [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62965) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1219.781884] env[62965]: DEBUG oslo_concurrency.lockutils [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.116s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1220.129007] env[62965]: DEBUG nova.network.neutron [None req-d57256ab-0308-4099-b0de-8a4778d9a1fd tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Updating instance_info_cache with network_info: [{"id": "416c6e70-3506-44e8-ab05-52aef45d2195", "address": "fa:16:3e:cb:c6:e5", "network": {"id": "4649fed7-c2ca-43a9-a183-7c99423da726", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1978698173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fab418e2fe1420793517663574b43bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43ad01d2-c7dd-453c-a929-8ad76294d13c", "external-id": "nsx-vlan-transportzone-176", "segmentation_id": 176, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap416c6e70-35", "ovs_interfaceid": "416c6e70-3506-44e8-ab05-52aef45d2195", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1220.632124] env[62965]: DEBUG oslo_concurrency.lockutils [None req-d57256ab-0308-4099-b0de-8a4778d9a1fd tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Releasing lock "refresh_cache-62b5855f-90f6-4480-bbbf-106e1ce6ae2d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1221.637961] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-d57256ab-0308-4099-b0de-8a4778d9a1fd tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1221.638414] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-09e3b6f7-ffb1-4e06-8610-765f126a89d8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.646163] env[62965]: DEBUG oslo_vmware.api [None req-d57256ab-0308-4099-b0de-8a4778d9a1fd tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1221.646163] env[62965]: value = "task-1390477" [ 1221.646163] env[62965]: _type = "Task" [ 1221.646163] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.653266] env[62965]: DEBUG oslo_vmware.api [None req-d57256ab-0308-4099-b0de-8a4778d9a1fd tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390477, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.156641] env[62965]: DEBUG oslo_vmware.api [None req-d57256ab-0308-4099-b0de-8a4778d9a1fd tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390477, 'name': PowerOnVM_Task, 'duration_secs': 0.349072} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.156935] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-d57256ab-0308-4099-b0de-8a4778d9a1fd tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1222.157157] env[62965]: DEBUG nova.compute.manager [None req-d57256ab-0308-4099-b0de-8a4778d9a1fd tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1222.157909] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-926467c9-f5c2-4b44-a9af-c23ba782419b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.318862] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-485b079c-c4ce-4fdd-97ba-fdf84d84e456 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.325881] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b0330d2b-abeb-4286-b7a4-09138088b147 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Suspending the VM {{(pid=62965) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1223.326119] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-ed122cef-d4e2-4acb-9f98-9be50629946c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.332037] env[62965]: DEBUG oslo_vmware.api [None req-b0330d2b-abeb-4286-b7a4-09138088b147 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1223.332037] env[62965]: value = "task-1390478" [ 1223.332037] env[62965]: _type = "Task" [ 1223.332037] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.339195] env[62965]: DEBUG oslo_vmware.api [None req-b0330d2b-abeb-4286-b7a4-09138088b147 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390478, 'name': SuspendVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.841375] env[62965]: DEBUG oslo_vmware.api [None req-b0330d2b-abeb-4286-b7a4-09138088b147 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390478, 'name': SuspendVM_Task} progress is 70%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.341843] env[62965]: DEBUG oslo_vmware.api [None req-b0330d2b-abeb-4286-b7a4-09138088b147 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390478, 'name': SuspendVM_Task, 'duration_secs': 0.817927} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.342217] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-b0330d2b-abeb-4286-b7a4-09138088b147 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Suspended the VM {{(pid=62965) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1224.342308] env[62965]: DEBUG nova.compute.manager [None req-b0330d2b-abeb-4286-b7a4-09138088b147 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1224.343032] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b90e68b-8382-48db-ad74-fee23ae4e0af {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.632346] env[62965]: INFO nova.compute.manager [None req-5fb10951-d320-4036-816f-d9cb6b4ad906 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Resuming [ 1225.632957] env[62965]: DEBUG nova.objects.instance [None req-5fb10951-d320-4036-816f-d9cb6b4ad906 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lazy-loading 'flavor' on Instance uuid 62b5855f-90f6-4480-bbbf-106e1ce6ae2d {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1226.640277] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5fb10951-d320-4036-816f-d9cb6b4ad906 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "refresh_cache-62b5855f-90f6-4480-bbbf-106e1ce6ae2d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1226.640668] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5fb10951-d320-4036-816f-d9cb6b4ad906 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquired lock "refresh_cache-62b5855f-90f6-4480-bbbf-106e1ce6ae2d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1226.640668] env[62965]: DEBUG nova.network.neutron [None req-5fb10951-d320-4036-816f-d9cb6b4ad906 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1227.338115] env[62965]: DEBUG nova.network.neutron [None req-5fb10951-d320-4036-816f-d9cb6b4ad906 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Updating instance_info_cache with network_info: [{"id": "416c6e70-3506-44e8-ab05-52aef45d2195", "address": "fa:16:3e:cb:c6:e5", "network": {"id": "4649fed7-c2ca-43a9-a183-7c99423da726", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1978698173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8fab418e2fe1420793517663574b43bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43ad01d2-c7dd-453c-a929-8ad76294d13c", "external-id": "nsx-vlan-transportzone-176", "segmentation_id": 176, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap416c6e70-35", "ovs_interfaceid": "416c6e70-3506-44e8-ab05-52aef45d2195", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1227.840592] env[62965]: DEBUG oslo_concurrency.lockutils [None req-5fb10951-d320-4036-816f-d9cb6b4ad906 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Releasing lock "refresh_cache-62b5855f-90f6-4480-bbbf-106e1ce6ae2d" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1227.841739] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c338c188-3e4d-4777-9899-677e353ddaf2 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.848571] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-5fb10951-d320-4036-816f-d9cb6b4ad906 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Resuming the VM {{(pid=62965) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1227.848797] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7c2d11c7-9462-4fd1-86a7-cfe1f043612e {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.854840] env[62965]: DEBUG oslo_vmware.api [None req-5fb10951-d320-4036-816f-d9cb6b4ad906 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1227.854840] env[62965]: value = "task-1390479" [ 1227.854840] env[62965]: _type = "Task" [ 1227.854840] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1227.862237] env[62965]: DEBUG oslo_vmware.api [None req-5fb10951-d320-4036-816f-d9cb6b4ad906 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390479, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.365250] env[62965]: DEBUG oslo_vmware.api [None req-5fb10951-d320-4036-816f-d9cb6b4ad906 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390479, 'name': PowerOnVM_Task, 'duration_secs': 0.493261} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.365522] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-5fb10951-d320-4036-816f-d9cb6b4ad906 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Resumed the VM {{(pid=62965) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1228.365760] env[62965]: DEBUG nova.compute.manager [None req-5fb10951-d320-4036-816f-d9cb6b4ad906 tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1228.366641] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-099b7440-4524-47d2-bc28-f0484d418a2c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.217485] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8ac8ae41-1cce-46a4-b0ba-51f25e6a087a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "62b5855f-90f6-4480-bbbf-106e1ce6ae2d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1229.217876] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8ac8ae41-1cce-46a4-b0ba-51f25e6a087a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "62b5855f-90f6-4480-bbbf-106e1ce6ae2d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1229.217990] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8ac8ae41-1cce-46a4-b0ba-51f25e6a087a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "62b5855f-90f6-4480-bbbf-106e1ce6ae2d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1229.218199] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8ac8ae41-1cce-46a4-b0ba-51f25e6a087a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "62b5855f-90f6-4480-bbbf-106e1ce6ae2d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1229.218375] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8ac8ae41-1cce-46a4-b0ba-51f25e6a087a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "62b5855f-90f6-4480-bbbf-106e1ce6ae2d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1229.220677] env[62965]: INFO nova.compute.manager [None req-8ac8ae41-1cce-46a4-b0ba-51f25e6a087a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Terminating instance [ 1229.725143] env[62965]: DEBUG nova.compute.manager [None req-8ac8ae41-1cce-46a4-b0ba-51f25e6a087a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1229.725380] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-8ac8ae41-1cce-46a4-b0ba-51f25e6a087a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1229.726378] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33c71d8a-e208-4f45-80b7-095ab8c685bf {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.733988] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ac8ae41-1cce-46a4-b0ba-51f25e6a087a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1229.734236] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2b6569e9-863c-4745-93b3-aa33c09634a9 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.740448] env[62965]: DEBUG oslo_vmware.api [None req-8ac8ae41-1cce-46a4-b0ba-51f25e6a087a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1229.740448] env[62965]: value = "task-1390480" [ 1229.740448] env[62965]: _type = "Task" [ 1229.740448] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1229.748429] env[62965]: DEBUG oslo_vmware.api [None req-8ac8ae41-1cce-46a4-b0ba-51f25e6a087a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390480, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.251369] env[62965]: DEBUG oslo_vmware.api [None req-8ac8ae41-1cce-46a4-b0ba-51f25e6a087a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390480, 'name': PowerOffVM_Task, 'duration_secs': 0.181989} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1230.251753] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ac8ae41-1cce-46a4-b0ba-51f25e6a087a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1230.251844] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-8ac8ae41-1cce-46a4-b0ba-51f25e6a087a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1230.252032] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-727b13ea-63ac-430a-aee0-8c5921735538 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.314685] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-8ac8ae41-1cce-46a4-b0ba-51f25e6a087a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1230.314920] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-8ac8ae41-1cce-46a4-b0ba-51f25e6a087a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Deleting contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1230.315099] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ac8ae41-1cce-46a4-b0ba-51f25e6a087a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Deleting the datastore file [datastore1] 62b5855f-90f6-4480-bbbf-106e1ce6ae2d {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1230.315388] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9c90e557-a16f-43ba-8102-6da3da698660 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.321748] env[62965]: DEBUG oslo_vmware.api [None req-8ac8ae41-1cce-46a4-b0ba-51f25e6a087a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for the task: (returnval){ [ 1230.321748] env[62965]: value = "task-1390482" [ 1230.321748] env[62965]: _type = "Task" [ 1230.321748] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1230.329215] env[62965]: DEBUG oslo_vmware.api [None req-8ac8ae41-1cce-46a4-b0ba-51f25e6a087a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390482, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.831198] env[62965]: DEBUG oslo_vmware.api [None req-8ac8ae41-1cce-46a4-b0ba-51f25e6a087a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Task: {'id': task-1390482, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.172215} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1230.831447] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-8ac8ae41-1cce-46a4-b0ba-51f25e6a087a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1230.831687] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-8ac8ae41-1cce-46a4-b0ba-51f25e6a087a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Deleted contents of the VM from datastore datastore1 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1230.831918] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-8ac8ae41-1cce-46a4-b0ba-51f25e6a087a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1230.832156] env[62965]: INFO nova.compute.manager [None req-8ac8ae41-1cce-46a4-b0ba-51f25e6a087a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1230.832444] env[62965]: DEBUG oslo.service.loopingcall [None req-8ac8ae41-1cce-46a4-b0ba-51f25e6a087a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1230.832669] env[62965]: DEBUG nova.compute.manager [-] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1230.832792] env[62965]: DEBUG nova.network.neutron [-] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1231.262576] env[62965]: DEBUG nova.compute.manager [req-64eaf9e2-2f42-4d9d-872c-5f9b68589f37 req-7970d838-bc04-4135-88d6-f83b0fd4ceeb service nova] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Received event network-vif-deleted-416c6e70-3506-44e8-ab05-52aef45d2195 {{(pid=62965) external_instance_event /opt/stack/nova/nova/compute/manager.py:11460}} [ 1231.262811] env[62965]: INFO nova.compute.manager [req-64eaf9e2-2f42-4d9d-872c-5f9b68589f37 req-7970d838-bc04-4135-88d6-f83b0fd4ceeb service nova] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Neutron deleted interface 416c6e70-3506-44e8-ab05-52aef45d2195; detaching it from the instance and deleting it from the info cache [ 1231.262945] env[62965]: DEBUG nova.network.neutron [req-64eaf9e2-2f42-4d9d-872c-5f9b68589f37 req-7970d838-bc04-4135-88d6-f83b0fd4ceeb service nova] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1231.742068] env[62965]: DEBUG nova.network.neutron [-] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1231.765270] env[62965]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c43e642b-aff9-4628-9445-cea38b9d9603 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.774243] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6cf3158-3bdf-451a-89a2-1b773c3fd3b8 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.796752] env[62965]: DEBUG nova.compute.manager [req-64eaf9e2-2f42-4d9d-872c-5f9b68589f37 req-7970d838-bc04-4135-88d6-f83b0fd4ceeb service nova] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Detach interface failed, port_id=416c6e70-3506-44e8-ab05-52aef45d2195, reason: Instance 62b5855f-90f6-4480-bbbf-106e1ce6ae2d could not be found. {{(pid=62965) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11294}} [ 1232.244421] env[62965]: INFO nova.compute.manager [-] [instance: 62b5855f-90f6-4480-bbbf-106e1ce6ae2d] Took 1.41 seconds to deallocate network for instance. [ 1232.751431] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8ac8ae41-1cce-46a4-b0ba-51f25e6a087a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1232.751804] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8ac8ae41-1cce-46a4-b0ba-51f25e6a087a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1232.751930] env[62965]: DEBUG nova.objects.instance [None req-8ac8ae41-1cce-46a4-b0ba-51f25e6a087a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lazy-loading 'resources' on Instance uuid 62b5855f-90f6-4480-bbbf-106e1ce6ae2d {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1233.286033] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf1ff204-fe38-46c1-842b-f874cf6eb019 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.293135] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fa8a4fd-967b-46c8-a370-c4f9f43822a6 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.323260] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-291f5a36-e792-4fe5-8b93-8e2ca0aa914a {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.329952] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-741c144c-4944-417d-8a1f-d404b381ef46 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.342325] env[62965]: DEBUG nova.compute.provider_tree [None req-8ac8ae41-1cce-46a4-b0ba-51f25e6a087a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1233.845522] env[62965]: DEBUG nova.scheduler.client.report [None req-8ac8ae41-1cce-46a4-b0ba-51f25e6a087a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1234.350136] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8ac8ae41-1cce-46a4-b0ba-51f25e6a087a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.598s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1234.372418] env[62965]: INFO nova.scheduler.client.report [None req-8ac8ae41-1cce-46a4-b0ba-51f25e6a087a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Deleted allocations for instance 62b5855f-90f6-4480-bbbf-106e1ce6ae2d [ 1234.881971] env[62965]: DEBUG oslo_concurrency.lockutils [None req-8ac8ae41-1cce-46a4-b0ba-51f25e6a087a tempest-ServerActionsTestJSON-1704521924 tempest-ServerActionsTestJSON-1704521924-project-member] Lock "62b5855f-90f6-4480-bbbf-106e1ce6ae2d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.664s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1238.088029] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Acquiring lock "f0211ba7-fe1f-4a87-a93f-11f1fff0addd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1238.088029] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Lock "f0211ba7-fe1f-4a87-a93f-11f1fff0addd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1238.591027] env[62965]: DEBUG nova.compute.manager [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Starting instance... {{(pid=62965) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1239.110394] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1239.110657] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1239.112215] env[62965]: INFO nova.compute.claims [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1240.145527] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f67b2ea9-67ad-43fe-af42-19039d3266a0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.153051] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d1afba2-6edf-45a7-b65a-ea9455446854 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.181698] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06dc60e4-7a01-4dd4-9a6a-208b919df8d7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.188471] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c6a4898-2bed-4078-9cbb-22046403e4ed {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.200833] env[62965]: DEBUG nova.compute.provider_tree [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1240.704360] env[62965]: DEBUG nova.scheduler.client.report [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1241.209600] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.099s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1241.210129] env[62965]: DEBUG nova.compute.manager [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Start building networks asynchronously for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1241.715845] env[62965]: DEBUG nova.compute.utils [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Using /dev/sd instead of None {{(pid=62965) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1241.717288] env[62965]: DEBUG nova.compute.manager [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Not allocating networking since 'none' was specified. {{(pid=62965) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 1242.219178] env[62965]: DEBUG nova.compute.manager [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Start building block device mappings for instance. {{(pid=62965) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1243.228582] env[62965]: DEBUG nova.compute.manager [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Start spawning the instance on the hypervisor. {{(pid=62965) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1243.254161] env[62965]: DEBUG nova.virt.hardware [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-31T12:29:22Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-31T12:29:05Z,direct_url=,disk_format='vmdk',id=523fe138-8131-44f9-96b3-87f923f60cb8,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='baa9ca2b16584dcd8d3b9eab9f9626eb',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-31T12:29:06Z,virtual_size=,visibility=), allow threads: False {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1243.254427] env[62965]: DEBUG nova.virt.hardware [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Flavor limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1243.254587] env[62965]: DEBUG nova.virt.hardware [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Image limits 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1243.254773] env[62965]: DEBUG nova.virt.hardware [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Flavor pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1243.254920] env[62965]: DEBUG nova.virt.hardware [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Image pref 0:0:0 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1243.255086] env[62965]: DEBUG nova.virt.hardware [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62965) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1243.255299] env[62965]: DEBUG nova.virt.hardware [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1243.255462] env[62965]: DEBUG nova.virt.hardware [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1243.255632] env[62965]: DEBUG nova.virt.hardware [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Got 1 possible topologies {{(pid=62965) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1243.255798] env[62965]: DEBUG nova.virt.hardware [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1243.255974] env[62965]: DEBUG nova.virt.hardware [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62965) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1243.256858] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46f41c75-fd38-423c-a763-7fb0d1c9b4f4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.264379] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04edaff9-5dbe-4159-8ced-23d9c78291a7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.277545] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Instance VIF info [] {{(pid=62965) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1243.282894] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Creating folder: Project (b46c8f903a1446f5b79d95d81a69e0d5). Parent ref: group-v294931. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1243.283158] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a0e32461-83d4-47d5-9a04-ff51f70553a1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.293101] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Created folder: Project (b46c8f903a1446f5b79d95d81a69e0d5) in parent group-v294931. [ 1243.293281] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Creating folder: Instances. Parent ref: group-v295073. {{(pid=62965) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1243.293484] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c55fd4b6-b882-4c4e-b73a-56359a196523 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.302605] env[62965]: INFO nova.virt.vmwareapi.vm_util [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Created folder: Instances in parent group-v295073. [ 1243.302820] env[62965]: DEBUG oslo.service.loopingcall [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1243.303008] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Creating VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1243.303210] env[62965]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7ef6a8d3-fb69-4dae-9633-fa581d697d78 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.318494] env[62965]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1243.318494] env[62965]: value = "task-1390485" [ 1243.318494] env[62965]: _type = "Task" [ 1243.318494] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1243.325289] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390485, 'name': CreateVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.828376] env[62965]: DEBUG oslo_vmware.api [-] Task: {'id': task-1390485, 'name': CreateVM_Task, 'duration_secs': 0.23782} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1243.828617] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Created VM on the ESX host {{(pid=62965) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1243.828975] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1243.829164] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1243.829509] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1243.829757] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa9fda3b-9eeb-4d60-8f2d-692228b275ca {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.834114] env[62965]: DEBUG oslo_vmware.api [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Waiting for the task: (returnval){ [ 1243.834114] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]528bcbfc-a60e-a271-27b9-287bd35621f4" [ 1243.834114] env[62965]: _type = "Task" [ 1243.834114] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1243.841199] env[62965]: DEBUG oslo_vmware.api [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]528bcbfc-a60e-a271-27b9-287bd35621f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.343987] env[62965]: DEBUG oslo_vmware.api [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]528bcbfc-a60e-a271-27b9-287bd35621f4, 'name': SearchDatastore_Task, 'duration_secs': 0.008945} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.344432] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1244.344567] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Processing image 523fe138-8131-44f9-96b3-87f923f60cb8 {{(pid=62965) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1244.344803] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1244.344967] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Acquired lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1244.345206] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1244.345464] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3d8929c0-73ec-4133-9636-e8893cb69c2f {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.353044] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62965) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1244.353225] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62965) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1244.353911] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b71f2d9b-3c04-4ab3-870a-aa911e26dd63 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.358882] env[62965]: DEBUG oslo_vmware.api [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Waiting for the task: (returnval){ [ 1244.358882] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52d21bd3-1cd5-2b0c-9215-af97089b1b39" [ 1244.358882] env[62965]: _type = "Task" [ 1244.358882] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.365676] env[62965]: DEBUG oslo_vmware.api [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52d21bd3-1cd5-2b0c-9215-af97089b1b39, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.869022] env[62965]: DEBUG oslo_vmware.api [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]52d21bd3-1cd5-2b0c-9215-af97089b1b39, 'name': SearchDatastore_Task, 'duration_secs': 0.007711} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.869782] env[62965]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3ae69ed-c766-41bd-a0c0-18b6f1ac5503 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.874303] env[62965]: DEBUG oslo_vmware.api [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Waiting for the task: (returnval){ [ 1244.874303] env[62965]: value = "session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5286d475-d6c2-3fb5-4832-0cfe7e4bda05" [ 1244.874303] env[62965]: _type = "Task" [ 1244.874303] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.881219] env[62965]: DEBUG oslo_vmware.api [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5286d475-d6c2-3fb5-4832-0cfe7e4bda05, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.386604] env[62965]: DEBUG oslo_vmware.api [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Task: {'id': session[52ce6cd3-b9d4-d664-c3fa-4cfb49aab8ab]5286d475-d6c2-3fb5-4832-0cfe7e4bda05, 'name': SearchDatastore_Task, 'duration_secs': 0.008438} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1245.386981] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Releasing lock "[datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1245.387131] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] f0211ba7-fe1f-4a87-a93f-11f1fff0addd/f0211ba7-fe1f-4a87-a93f-11f1fff0addd.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1245.387395] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8bd93466-364b-4306-b863-353ca63a2fd3 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.395037] env[62965]: DEBUG oslo_vmware.api [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Waiting for the task: (returnval){ [ 1245.395037] env[62965]: value = "task-1390486" [ 1245.395037] env[62965]: _type = "Task" [ 1245.395037] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1245.402055] env[62965]: DEBUG oslo_vmware.api [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Task: {'id': task-1390486, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.904660] env[62965]: DEBUG oslo_vmware.api [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Task: {'id': task-1390486, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.44002} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1245.904948] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/523fe138-8131-44f9-96b3-87f923f60cb8/523fe138-8131-44f9-96b3-87f923f60cb8.vmdk to [datastore2] f0211ba7-fe1f-4a87-a93f-11f1fff0addd/f0211ba7-fe1f-4a87-a93f-11f1fff0addd.vmdk {{(pid=62965) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1245.906603] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Extending root virtual disk to 1048576 {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1245.906603] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b0aa2b20-e1c8-4389-af1e-a45c2f971499 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.912404] env[62965]: DEBUG oslo_vmware.api [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Waiting for the task: (returnval){ [ 1245.912404] env[62965]: value = "task-1390487" [ 1245.912404] env[62965]: _type = "Task" [ 1245.912404] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1245.919279] env[62965]: DEBUG oslo_vmware.api [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Task: {'id': task-1390487, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.421931] env[62965]: DEBUG oslo_vmware.api [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Task: {'id': task-1390487, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066202} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1246.422247] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Extended root virtual disk {{(pid=62965) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1246.422965] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e59c7622-8734-4148-928c-0d793443b316 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.441772] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] f0211ba7-fe1f-4a87-a93f-11f1fff0addd/f0211ba7-fe1f-4a87-a93f-11f1fff0addd.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1246.441949] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-261b6479-9f3f-4a36-ab61-a3c8261d1d5c {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.460751] env[62965]: DEBUG oslo_vmware.api [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Waiting for the task: (returnval){ [ 1246.460751] env[62965]: value = "task-1390488" [ 1246.460751] env[62965]: _type = "Task" [ 1246.460751] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1246.467803] env[62965]: DEBUG oslo_vmware.api [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Task: {'id': task-1390488, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.971012] env[62965]: DEBUG oslo_vmware.api [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Task: {'id': task-1390488, 'name': ReconfigVM_Task, 'duration_secs': 0.259203} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1246.971163] env[62965]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Reconfigured VM instance instance-0000006f to attach disk [datastore2] f0211ba7-fe1f-4a87-a93f-11f1fff0addd/f0211ba7-fe1f-4a87-a93f-11f1fff0addd.vmdk or device None with type sparse {{(pid=62965) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1246.971763] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5e21acd1-f8f9-4873-8660-e5c815ac5053 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.978578] env[62965]: DEBUG oslo_vmware.api [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Waiting for the task: (returnval){ [ 1246.978578] env[62965]: value = "task-1390489" [ 1246.978578] env[62965]: _type = "Task" [ 1246.978578] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1246.985754] env[62965]: DEBUG oslo_vmware.api [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Task: {'id': task-1390489, 'name': Rename_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1247.488341] env[62965]: DEBUG oslo_vmware.api [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Task: {'id': task-1390489, 'name': Rename_Task, 'duration_secs': 0.124532} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1247.488624] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Powering on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1247.488857] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8a31e4c4-8eb2-4c8d-8cb0-f2ec76d52dbb {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.494566] env[62965]: DEBUG oslo_vmware.api [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Waiting for the task: (returnval){ [ 1247.494566] env[62965]: value = "task-1390490" [ 1247.494566] env[62965]: _type = "Task" [ 1247.494566] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1247.501446] env[62965]: DEBUG oslo_vmware.api [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Task: {'id': task-1390490, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1248.004237] env[62965]: DEBUG oslo_vmware.api [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Task: {'id': task-1390490, 'name': PowerOnVM_Task, 'duration_secs': 0.389893} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1248.004511] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Powered on the VM {{(pid=62965) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1248.004711] env[62965]: INFO nova.compute.manager [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Took 4.78 seconds to spawn the instance on the hypervisor. [ 1248.004930] env[62965]: DEBUG nova.compute.manager [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1248.005667] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38eed30c-fc97-4d80-94fb-c0f57aa222a0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.523692] env[62965]: INFO nova.compute.manager [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Took 9.43 seconds to build instance. [ 1249.026062] env[62965]: DEBUG oslo_concurrency.lockutils [None req-e5db2c95-d28e-4f74-b5ce-8f4b8071a7dd tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Lock "f0211ba7-fe1f-4a87-a93f-11f1fff0addd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 10.938s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1249.130859] env[62965]: DEBUG nova.compute.manager [None req-ba3be42f-28ca-4587-8156-ebf7155ccda0 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Checking state {{(pid=62965) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1249.131828] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f15cfb04-4bce-4793-9162-567b7d7a9c39 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.200643] env[62965]: DEBUG oslo_concurrency.lockutils [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Acquiring lock "f0211ba7-fe1f-4a87-a93f-11f1fff0addd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1249.200882] env[62965]: DEBUG oslo_concurrency.lockutils [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Lock "f0211ba7-fe1f-4a87-a93f-11f1fff0addd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1249.201113] env[62965]: DEBUG oslo_concurrency.lockutils [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Acquiring lock "f0211ba7-fe1f-4a87-a93f-11f1fff0addd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1249.201616] env[62965]: DEBUG oslo_concurrency.lockutils [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Lock "f0211ba7-fe1f-4a87-a93f-11f1fff0addd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1249.201820] env[62965]: DEBUG oslo_concurrency.lockutils [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Lock "f0211ba7-fe1f-4a87-a93f-11f1fff0addd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1249.203714] env[62965]: INFO nova.compute.manager [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Terminating instance [ 1249.641965] env[62965]: INFO nova.compute.manager [None req-ba3be42f-28ca-4587-8156-ebf7155ccda0 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] instance snapshotting [ 1249.642605] env[62965]: DEBUG nova.objects.instance [None req-ba3be42f-28ca-4587-8156-ebf7155ccda0 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Lazy-loading 'flavor' on Instance uuid f0211ba7-fe1f-4a87-a93f-11f1fff0addd {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1249.707687] env[62965]: DEBUG oslo_concurrency.lockutils [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Acquiring lock "refresh_cache-f0211ba7-fe1f-4a87-a93f-11f1fff0addd" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1249.707894] env[62965]: DEBUG oslo_concurrency.lockutils [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Acquired lock "refresh_cache-f0211ba7-fe1f-4a87-a93f-11f1fff0addd" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1249.708062] env[62965]: DEBUG nova.network.neutron [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Building network info cache for instance {{(pid=62965) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1250.148508] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a72e2348-7ca5-430b-afd1-b056799a49a0 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.164920] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0ce2e91-e4b9-4704-8abd-b076fe1e1e51 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.226272] env[62965]: DEBUG nova.network.neutron [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1250.284243] env[62965]: DEBUG nova.network.neutron [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1250.674337] env[62965]: DEBUG nova.compute.manager [None req-ba3be42f-28ca-4587-8156-ebf7155ccda0 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Instance disappeared during snapshot {{(pid=62965) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4580}} [ 1250.787575] env[62965]: DEBUG oslo_concurrency.lockutils [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Releasing lock "refresh_cache-f0211ba7-fe1f-4a87-a93f-11f1fff0addd" {{(pid=62965) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1250.788023] env[62965]: DEBUG nova.compute.manager [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Start destroying the instance on the hypervisor. {{(pid=62965) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1250.788294] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Destroying instance {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1250.789184] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c45b7175-7b19-455d-a2d9-d13db775ffd7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.794267] env[62965]: DEBUG nova.compute.manager [None req-ba3be42f-28ca-4587-8156-ebf7155ccda0 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Found 0 images (rotation: 2) {{(pid=62965) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4883}} [ 1250.797607] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Powering off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1250.798055] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b08ef6ef-ffc7-4c23-8ba8-757c45de2518 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.805381] env[62965]: DEBUG oslo_vmware.api [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Waiting for the task: (returnval){ [ 1250.805381] env[62965]: value = "task-1390491" [ 1250.805381] env[62965]: _type = "Task" [ 1250.805381] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1250.815241] env[62965]: DEBUG oslo_vmware.api [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Task: {'id': task-1390491, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1251.315629] env[62965]: DEBUG oslo_vmware.api [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Task: {'id': task-1390491, 'name': PowerOffVM_Task, 'duration_secs': 0.184634} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1251.315881] env[62965]: DEBUG nova.virt.vmwareapi.vm_util [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Powered off the VM {{(pid=62965) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1251.316059] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Unregistering the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1251.316311] env[62965]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c1b1534a-c6e6-4792-98d8-00bda3230459 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.340799] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Unregistered the VM {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1251.341038] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Deleting contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1251.341187] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Deleting the datastore file [datastore2] f0211ba7-fe1f-4a87-a93f-11f1fff0addd {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1251.341438] env[62965]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3cea940a-6d8f-488a-a55f-9d8c4c9a0f17 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.347480] env[62965]: DEBUG oslo_vmware.api [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Waiting for the task: (returnval){ [ 1251.347480] env[62965]: value = "task-1390493" [ 1251.347480] env[62965]: _type = "Task" [ 1251.347480] env[62965]: } to complete. {{(pid=62965) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1251.354601] env[62965]: DEBUG oslo_vmware.api [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Task: {'id': task-1390493, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1251.859522] env[62965]: DEBUG oslo_vmware.api [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Task: {'id': task-1390493, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.085887} completed successfully. {{(pid=62965) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1251.859852] env[62965]: DEBUG nova.virt.vmwareapi.ds_util [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Deleted the datastore file {{(pid=62965) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1251.859983] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Deleted contents of the VM from datastore datastore2 {{(pid=62965) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1251.860175] env[62965]: DEBUG nova.virt.vmwareapi.vmops [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Instance destroyed {{(pid=62965) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1251.860362] env[62965]: INFO nova.compute.manager [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Took 1.07 seconds to destroy the instance on the hypervisor. [ 1251.860609] env[62965]: DEBUG oslo.service.loopingcall [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62965) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1251.860815] env[62965]: DEBUG nova.compute.manager [-] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Deallocating network for instance {{(pid=62965) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1251.860902] env[62965]: DEBUG nova.network.neutron [-] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] deallocate_for_instance() {{(pid=62965) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1251.878134] env[62965]: DEBUG nova.network.neutron [-] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Instance cache missing network info. {{(pid=62965) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1252.381058] env[62965]: DEBUG nova.network.neutron [-] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Updating instance_info_cache with network_info: [] {{(pid=62965) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1252.882966] env[62965]: INFO nova.compute.manager [-] [instance: f0211ba7-fe1f-4a87-a93f-11f1fff0addd] Took 1.02 seconds to deallocate network for instance. [ 1253.390264] env[62965]: DEBUG oslo_concurrency.lockutils [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1253.390559] env[62965]: DEBUG oslo_concurrency.lockutils [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1253.390742] env[62965]: DEBUG nova.objects.instance [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Lazy-loading 'resources' on Instance uuid f0211ba7-fe1f-4a87-a93f-11f1fff0addd {{(pid=62965) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1253.926961] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4740b1e-8c24-4cd9-9198-a68c4035adf1 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.935643] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-651e17aa-0d7e-4a44-abf3-a8c17815aaf4 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.964222] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdd6f171-4112-4d17-8e24-a09698c638e7 {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.971041] env[62965]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b61616b1-db31-456f-80e8-69aee20c2b8b {{(pid=62965) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.984567] env[62965]: DEBUG nova.compute.provider_tree [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Inventory has not changed in ProviderTree for provider: 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 {{(pid=62965) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1254.487376] env[62965]: DEBUG nova.scheduler.client.report [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Inventory has not changed for provider 20a9cd55-63ed-4d2c-aab8-50c3d9cf92e8 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 175, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62965) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1254.992588] env[62965]: DEBUG oslo_concurrency.lockutils [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.602s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1255.011209] env[62965]: INFO nova.scheduler.client.report [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Deleted allocations for instance f0211ba7-fe1f-4a87-a93f-11f1fff0addd [ 1255.519316] env[62965]: DEBUG oslo_concurrency.lockutils [None req-564a0227-34ee-494d-ae96-c9aff9a5f666 tempest-ServersAaction247Test-244575627 tempest-ServersAaction247Test-244575627-project-member] Lock "f0211ba7-fe1f-4a87-a93f-11f1fff0addd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.318s {{(pid=62965) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1272.782544] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1272.782908] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Starting heal instance info cache {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10257}} [ 1272.782908] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Rebuilding the list of instances to heal {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10261}} [ 1273.286558] env[62965]: DEBUG nova.compute.manager [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Didn't find any instances for network info cache update. {{(pid=62965) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10343}} [ 1273.286784] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1273.286941] env[62965]: DEBUG oslo_service.periodic_task [None req-ec6b6257-ecd3-4ce2-aab8-34545072d468 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62965) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}}